diff --git "a/log/debug_0.log" "b/log/debug_0.log" --- "a/log/debug_0.log" +++ "b/log/debug_0.log" @@ -14044,3 +14044,2009 @@ Use FP16 precision: False 01/27/2022 09:00:53 - INFO - codeparrot_training - Step 13998: {'lr': 0.00042679983379736324, 'samples': 2687808, 'steps': 13998, 'loss/train': 1.0843375325202942} 01/27/2022 09:00:56 - INFO - codeparrot_training - Step 13999: {'lr': 0.0004267882649256525, 'samples': 2688000, 'steps': 13999, 'loss/train': 1.0735023021697998} 01/27/2022 09:00:56 - INFO - codeparrot_training - Evaluating and saving model checkpoint +01/27/2022 09:01:14 - WARNING - huggingface_hub.repository - Several commits (7) will be pushed upstream. +01/27/2022 09:01:14 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. +01/27/2022 09:01:48 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/codeparrot-neo-125M-py + 2f6074d..283559f royal-monkey-12 -> royal-monkey-12 + +01/27/2022 09:01:53 - INFO - codeparrot_training - Step 14000: {'lr': 0.00042677669529663686, 'samples': 2688192, 'steps': 14000, 'loss/train': 0.9348163604736328} +01/27/2022 09:01:56 - INFO - codeparrot_training - Step 14001: {'lr': 0.0004267651249103661, 'samples': 2688384, 'steps': 14001, 'loss/train': 0.03093577455729246} +01/27/2022 09:01:59 - INFO - codeparrot_training - Step 14002: {'lr': 0.00042675355376688964, 'samples': 2688576, 'steps': 14002, 'loss/train': 0.6535166501998901} +01/27/2022 09:02:02 - INFO - codeparrot_training - Step 14003: {'lr': 0.000426741981866257, 'samples': 2688768, 'steps': 14003, 'loss/train': 1.2150499820709229} +01/27/2022 09:02:07 - INFO - codeparrot_training - Step 14004: {'lr': 0.00042673040920851793, 'samples': 2688960, 'steps': 14004, 'loss/train': 0.5038377642631531} +01/27/2022 09:02:10 - INFO - codeparrot_training - Step 14005: {'lr': 0.00042671883579372186, 'samples': 2689152, 'steps': 14005, 'loss/train': 0.7759323120117188} +01/27/2022 09:02:13 - INFO - codeparrot_training - Step 14006: {'lr': 0.00042670726162191843, 'samples': 2689344, 'steps': 14006, 'loss/train': 0.9634372293949127} +01/27/2022 09:02:16 - INFO - codeparrot_training - Step 14007: {'lr': 0.0004266956866931572, 'samples': 2689536, 'steps': 14007, 'loss/train': 0.7067593485116959} +01/27/2022 09:02:19 - INFO - codeparrot_training - Step 14008: {'lr': 0.0004266841110074878, 'samples': 2689728, 'steps': 14008, 'loss/train': 0.24828940629959106} +01/27/2022 09:02:22 - INFO - codeparrot_training - Step 14009: {'lr': 0.0004266725345649597, 'samples': 2689920, 'steps': 14009, 'loss/train': 0.22894486784934998} +01/27/2022 09:02:26 - INFO - codeparrot_training - Step 14010: {'lr': 0.0004266609573656226, 'samples': 2690112, 'steps': 14010, 'loss/train': 1.2506467401981354} +01/27/2022 09:02:29 - INFO - codeparrot_training - Step 14011: {'lr': 0.000426649379409526, 'samples': 2690304, 'steps': 14011, 'loss/train': 1.0491205751895905} +01/27/2022 09:02:32 - INFO - codeparrot_training - Step 14012: {'lr': 0.00042663780069671965, 'samples': 2690496, 'steps': 14012, 'loss/train': 1.26598459482193} +01/27/2022 09:02:36 - INFO - codeparrot_training - Step 14013: {'lr': 0.000426626221227253, 'samples': 2690688, 'steps': 14013, 'loss/train': 0.3706120401620865} +01/27/2022 09:02:39 - INFO - codeparrot_training - Step 14014: {'lr': 0.00042661464100117566, 'samples': 2690880, 'steps': 14014, 'loss/train': 0.7020441144704819} +01/27/2022 09:02:43 - INFO - codeparrot_training - Step 14015: {'lr': 0.00042660306001853735, 'samples': 2691072, 'steps': 14015, 'loss/train': 0.5079659074544907} +01/27/2022 09:02:46 - INFO - codeparrot_training - Step 14016: {'lr': 0.0004265914782793875, 'samples': 2691264, 'steps': 14016, 'loss/train': 1.1829104125499725} +01/27/2022 09:02:49 - INFO - codeparrot_training - Step 14017: {'lr': 0.000426579895783776, 'samples': 2691456, 'steps': 14017, 'loss/train': 0.7406394481658936} +01/27/2022 09:02:52 - INFO - codeparrot_training - Step 14018: {'lr': 0.0004265683125317521, 'samples': 2691648, 'steps': 14018, 'loss/train': 0.7921828329563141} +01/27/2022 09:02:55 - INFO - codeparrot_training - Step 14019: {'lr': 0.0004265567285233658, 'samples': 2691840, 'steps': 14019, 'loss/train': 0.6530756950378418} +01/27/2022 09:02:58 - INFO - codeparrot_training - Step 14020: {'lr': 0.0004265451437586664, 'samples': 2692032, 'steps': 14020, 'loss/train': 0.7967126369476318} +01/27/2022 09:03:01 - INFO - codeparrot_training - Step 14021: {'lr': 0.0004265335582377038, 'samples': 2692224, 'steps': 14021, 'loss/train': 0.9161058962345123} +01/27/2022 09:03:09 - INFO - codeparrot_training - Step 14022: {'lr': 0.0004265219719605273, 'samples': 2692416, 'steps': 14022, 'loss/train': 0.8824150264263153} +01/27/2022 09:03:12 - INFO - codeparrot_training - Step 14023: {'lr': 0.0004265103849271869, 'samples': 2692608, 'steps': 14023, 'loss/train': 0.7025805115699768} +01/27/2022 09:03:15 - INFO - codeparrot_training - Step 14024: {'lr': 0.000426498797137732, 'samples': 2692800, 'steps': 14024, 'loss/train': 0.745609849691391} +01/27/2022 09:03:18 - INFO - codeparrot_training - Step 14025: {'lr': 0.0004264872085922122, 'samples': 2692992, 'steps': 14025, 'loss/train': 0.7472414821386337} +01/27/2022 09:03:21 - INFO - codeparrot_training - Step 14026: {'lr': 0.0004264756192906774, 'samples': 2693184, 'steps': 14026, 'loss/train': 0.8370361626148224} +01/27/2022 09:03:24 - INFO - codeparrot_training - Step 14027: {'lr': 0.000426464029233177, 'samples': 2693376, 'steps': 14027, 'loss/train': 0.4303290992975235} +01/27/2022 09:03:28 - INFO - codeparrot_training - Step 14028: {'lr': 0.0004264524384197608, 'samples': 2693568, 'steps': 14028, 'loss/train': 0.9047146439552307} +01/27/2022 09:03:31 - INFO - codeparrot_training - Step 14029: {'lr': 0.0004264408468504783, 'samples': 2693760, 'steps': 14029, 'loss/train': 0.8816965520381927} +01/27/2022 09:03:35 - INFO - codeparrot_training - Step 14030: {'lr': 0.00042642925452537927, 'samples': 2693952, 'steps': 14030, 'loss/train': 1.0178795456886292} +01/27/2022 09:03:38 - INFO - codeparrot_training - Step 14031: {'lr': 0.0004264176614445133, 'samples': 2694144, 'steps': 14031, 'loss/train': 0.6776069104671478} +01/27/2022 09:03:41 - INFO - codeparrot_training - Step 14032: {'lr': 0.0004264060676079302, 'samples': 2694336, 'steps': 14032, 'loss/train': 0.3680715411901474} +01/27/2022 09:03:45 - INFO - codeparrot_training - Step 14033: {'lr': 0.00042639447301567944, 'samples': 2694528, 'steps': 14033, 'loss/train': 0.47217366099357605} +01/27/2022 09:03:48 - INFO - codeparrot_training - Step 14034: {'lr': 0.0004263828776678108, 'samples': 2694720, 'steps': 14034, 'loss/train': 1.163200557231903} +01/27/2022 09:03:51 - INFO - codeparrot_training - Step 14035: {'lr': 0.00042637128156437385, 'samples': 2694912, 'steps': 14035, 'loss/train': 1.0121871829032898} +01/27/2022 09:03:54 - INFO - codeparrot_training - Step 14036: {'lr': 0.0004263596847054184, 'samples': 2695104, 'steps': 14036, 'loss/train': 0.6059588491916656} +01/27/2022 09:03:57 - INFO - codeparrot_training - Step 14037: {'lr': 0.00042634808709099403, 'samples': 2695296, 'steps': 14037, 'loss/train': 0.9278047978878021} +01/27/2022 09:04:00 - INFO - codeparrot_training - Step 14038: {'lr': 0.0004263364887211505, 'samples': 2695488, 'steps': 14038, 'loss/train': 1.0545045733451843} +01/27/2022 09:04:05 - INFO - codeparrot_training - Step 14039: {'lr': 0.0004263248895959374, 'samples': 2695680, 'steps': 14039, 'loss/train': 0.9491678774356842} +01/27/2022 09:04:08 - INFO - codeparrot_training - Step 14040: {'lr': 0.0004263132897154044, 'samples': 2695872, 'steps': 14040, 'loss/train': 0.6365851163864136} +01/27/2022 09:04:11 - INFO - codeparrot_training - Step 14041: {'lr': 0.0004263016890796014, 'samples': 2696064, 'steps': 14041, 'loss/train': 0.9167093932628632} +01/27/2022 09:04:14 - INFO - codeparrot_training - Step 14042: {'lr': 0.0004262900876885778, 'samples': 2696256, 'steps': 14042, 'loss/train': 0.5883894860744476} +01/27/2022 09:04:17 - INFO - codeparrot_training - Step 14043: {'lr': 0.0004262784855423836, 'samples': 2696448, 'steps': 14043, 'loss/train': 0.7767824828624725} +01/27/2022 09:04:20 - INFO - codeparrot_training - Step 14044: {'lr': 0.00042626688264106816, 'samples': 2696640, 'steps': 14044, 'loss/train': 1.11994868516922} +01/27/2022 09:04:24 - INFO - codeparrot_training - Step 14045: {'lr': 0.00042625527898468155, 'samples': 2696832, 'steps': 14045, 'loss/train': 0.4896458834409714} +01/27/2022 09:04:27 - INFO - codeparrot_training - Step 14046: {'lr': 0.0004262436745732732, 'samples': 2697024, 'steps': 14046, 'loss/train': 1.068556934595108} +01/27/2022 09:04:30 - INFO - codeparrot_training - Step 14047: {'lr': 0.00042623206940689285, 'samples': 2697216, 'steps': 14047, 'loss/train': 0.8149688243865967} +01/27/2022 09:04:36 - INFO - codeparrot_training - Step 14048: {'lr': 0.00042622046348559034, 'samples': 2697408, 'steps': 14048, 'loss/train': 0.7361974865198135} +01/27/2022 09:04:39 - INFO - codeparrot_training - Step 14049: {'lr': 0.0004262088568094153, 'samples': 2697600, 'steps': 14049, 'loss/train': 1.253288984298706} +01/27/2022 09:04:42 - INFO - codeparrot_training - Step 14050: {'lr': 0.0004261972493784175, 'samples': 2697792, 'steps': 14050, 'loss/train': 1.1130180358886719} +01/27/2022 09:04:46 - INFO - codeparrot_training - Step 14051: {'lr': 0.0004261856411926467, 'samples': 2697984, 'steps': 14051, 'loss/train': 0.9638851583003998} +01/27/2022 09:04:49 - INFO - codeparrot_training - Step 14052: {'lr': 0.0004261740322521525, 'samples': 2698176, 'steps': 14052, 'loss/train': 0.5950068533420563} +01/27/2022 09:04:52 - INFO - codeparrot_training - Step 14053: {'lr': 0.00042616242255698463, 'samples': 2698368, 'steps': 14053, 'loss/train': 0.8018048107624054} +01/27/2022 09:04:55 - INFO - codeparrot_training - Step 14054: {'lr': 0.0004261508121071929, 'samples': 2698560, 'steps': 14054, 'loss/train': 0.7273528575897217} +01/27/2022 09:04:58 - INFO - codeparrot_training - Step 14055: {'lr': 0.00042613920090282706, 'samples': 2698752, 'steps': 14055, 'loss/train': 0.9404013454914093} +01/27/2022 09:05:01 - INFO - codeparrot_training - Step 14056: {'lr': 0.0004261275889439368, 'samples': 2698944, 'steps': 14056, 'loss/train': 0.32792188972234726} +01/27/2022 09:05:06 - INFO - codeparrot_training - Step 14057: {'lr': 0.0004261159762305719, 'samples': 2699136, 'steps': 14057, 'loss/train': 0.8232925236225128} +01/27/2022 09:05:09 - INFO - codeparrot_training - Step 14058: {'lr': 0.00042610436276278196, 'samples': 2699328, 'steps': 14058, 'loss/train': 0.8409935832023621} +01/27/2022 09:05:12 - INFO - codeparrot_training - Step 14059: {'lr': 0.00042609274854061695, 'samples': 2699520, 'steps': 14059, 'loss/train': 1.1656158864498138} +01/27/2022 09:05:15 - INFO - codeparrot_training - Step 14060: {'lr': 0.0004260811335641266, 'samples': 2699712, 'steps': 14060, 'loss/train': 0.491950660943985} +01/27/2022 09:05:18 - INFO - codeparrot_training - Step 14061: {'lr': 0.00042606951783336045, 'samples': 2699904, 'steps': 14061, 'loss/train': 0.05803181603550911} +01/27/2022 09:05:22 - INFO - codeparrot_training - Step 14062: {'lr': 0.0004260579013483684, 'samples': 2700096, 'steps': 14062, 'loss/train': 0.9512220919132233} +01/27/2022 09:05:25 - INFO - codeparrot_training - Step 14063: {'lr': 0.0004260462841092003, 'samples': 2700288, 'steps': 14063, 'loss/train': 1.027407556772232} +01/27/2022 09:05:28 - INFO - codeparrot_training - Step 14064: {'lr': 0.00042603466611590575, 'samples': 2700480, 'steps': 14064, 'loss/train': 0.7963347136974335} +01/27/2022 09:05:31 - INFO - codeparrot_training - Step 14065: {'lr': 0.00042602304736853464, 'samples': 2700672, 'steps': 14065, 'loss/train': 0.7517253756523132} +01/27/2022 09:05:36 - INFO - codeparrot_training - Step 14066: {'lr': 0.00042601142786713664, 'samples': 2700864, 'steps': 14066, 'loss/train': 1.1680482029914856} +01/27/2022 09:05:39 - INFO - codeparrot_training - Step 14067: {'lr': 0.0004259998076117616, 'samples': 2701056, 'steps': 14067, 'loss/train': 1.5107131004333496} +01/27/2022 09:05:42 - INFO - codeparrot_training - Step 14068: {'lr': 0.00042598818660245926, 'samples': 2701248, 'steps': 14068, 'loss/train': 1.087300032377243} +01/27/2022 09:05:45 - INFO - codeparrot_training - Step 14069: {'lr': 0.00042597656483927936, 'samples': 2701440, 'steps': 14069, 'loss/train': 0.4929453134536743} +01/27/2022 09:05:48 - INFO - codeparrot_training - Step 14070: {'lr': 0.0004259649423222718, 'samples': 2701632, 'steps': 14070, 'loss/train': 0.45052555203437805} +01/27/2022 09:05:52 - INFO - codeparrot_training - Step 14071: {'lr': 0.0004259533190514863, 'samples': 2701824, 'steps': 14071, 'loss/train': 1.63324373960495} +01/27/2022 09:05:55 - INFO - codeparrot_training - Step 14072: {'lr': 0.00042594169502697265, 'samples': 2702016, 'steps': 14072, 'loss/train': 0.3566892296075821} +01/27/2022 09:05:58 - INFO - codeparrot_training - Step 14073: {'lr': 0.0004259300702487806, 'samples': 2702208, 'steps': 14073, 'loss/train': 0.7061193734407425} +01/27/2022 09:06:01 - INFO - codeparrot_training - Step 14074: {'lr': 0.00042591844471696005, 'samples': 2702400, 'steps': 14074, 'loss/train': 0.7390531897544861} +01/27/2022 09:06:07 - INFO - codeparrot_training - Step 14075: {'lr': 0.00042590681843156073, 'samples': 2702592, 'steps': 14075, 'loss/train': 0.8177030682563782} +01/27/2022 09:06:11 - INFO - codeparrot_training - Step 14076: {'lr': 0.00042589519139263246, 'samples': 2702784, 'steps': 14076, 'loss/train': 1.420966923236847} +01/27/2022 09:06:14 - INFO - codeparrot_training - Step 14077: {'lr': 0.0004258835636002251, 'samples': 2702976, 'steps': 14077, 'loss/train': 0.6224155426025391} +01/27/2022 09:06:17 - INFO - codeparrot_training - Step 14078: {'lr': 0.0004258719350543883, 'samples': 2703168, 'steps': 14078, 'loss/train': 0.9374057650566101} +01/27/2022 09:06:20 - INFO - codeparrot_training - Step 14079: {'lr': 0.00042586030575517196, 'samples': 2703360, 'steps': 14079, 'loss/train': 0.5243774056434631} +01/27/2022 09:06:23 - INFO - codeparrot_training - Step 14080: {'lr': 0.00042584867570262595, 'samples': 2703552, 'steps': 14080, 'loss/train': 1.0744171142578125} +01/27/2022 09:06:26 - INFO - codeparrot_training - Step 14081: {'lr': 0.00042583704489680007, 'samples': 2703744, 'steps': 14081, 'loss/train': 0.7812465727329254} +01/27/2022 09:06:29 - INFO - codeparrot_training - Step 14082: {'lr': 0.00042582541333774414, 'samples': 2703936, 'steps': 14082, 'loss/train': 0.7918497025966644} +01/27/2022 09:06:33 - INFO - codeparrot_training - Step 14083: {'lr': 0.0004258137810255079, 'samples': 2704128, 'steps': 14083, 'loss/train': 0.757999062538147} +01/27/2022 09:06:37 - INFO - codeparrot_training - Step 14084: {'lr': 0.0004258021479601414, 'samples': 2704320, 'steps': 14084, 'loss/train': 0.809131532907486} +01/27/2022 09:06:41 - INFO - codeparrot_training - Step 14085: {'lr': 0.00042579051414169417, 'samples': 2704512, 'steps': 14085, 'loss/train': 0.7772054672241211} +01/27/2022 09:06:44 - INFO - codeparrot_training - Step 14086: {'lr': 0.0004257788795702162, 'samples': 2704704, 'steps': 14086, 'loss/train': 0.2788681089878082} +01/27/2022 09:06:47 - INFO - codeparrot_training - Step 14087: {'lr': 0.0004257672442457574, 'samples': 2704896, 'steps': 14087, 'loss/train': 0.27799395471811295} +01/27/2022 09:06:50 - INFO - codeparrot_training - Step 14088: {'lr': 0.00042575560816836755, 'samples': 2705088, 'steps': 14088, 'loss/train': 0.875478208065033} +01/27/2022 09:06:53 - INFO - codeparrot_training - Step 14089: {'lr': 0.00042574397133809646, 'samples': 2705280, 'steps': 14089, 'loss/train': 0.924744039773941} +01/27/2022 09:06:56 - INFO - codeparrot_training - Step 14090: {'lr': 0.000425732333754994, 'samples': 2705472, 'steps': 14090, 'loss/train': 0.07110870257019997} +01/27/2022 09:07:00 - INFO - codeparrot_training - Step 14091: {'lr': 0.00042572069541911, 'samples': 2705664, 'steps': 14091, 'loss/train': 0.8246610760688782} +01/27/2022 09:07:03 - INFO - codeparrot_training - Step 14092: {'lr': 0.0004257090563304943, 'samples': 2705856, 'steps': 14092, 'loss/train': 1.0051999390125275} +01/27/2022 09:07:09 - INFO - codeparrot_training - Step 14093: {'lr': 0.0004256974164891969, 'samples': 2706048, 'steps': 14093, 'loss/train': 0.6540942192077637} +01/27/2022 09:07:12 - INFO - codeparrot_training - Step 14094: {'lr': 0.00042568577589526744, 'samples': 2706240, 'steps': 14094, 'loss/train': 1.091096431016922} +01/27/2022 09:07:15 - INFO - codeparrot_training - Step 14095: {'lr': 0.00042567413454875605, 'samples': 2706432, 'steps': 14095, 'loss/train': 0.908374696969986} +01/27/2022 09:07:19 - INFO - codeparrot_training - Step 14096: {'lr': 0.00042566249244971235, 'samples': 2706624, 'steps': 14096, 'loss/train': 0.5112421810626984} +01/27/2022 09:07:22 - INFO - codeparrot_training - Step 14097: {'lr': 0.0004256508495981863, 'samples': 2706816, 'steps': 14097, 'loss/train': 0.5233426541090012} +01/27/2022 09:07:25 - INFO - codeparrot_training - Step 14098: {'lr': 0.00042563920599422776, 'samples': 2707008, 'steps': 14098, 'loss/train': 1.1279726922512054} +01/27/2022 09:07:28 - INFO - codeparrot_training - Step 14099: {'lr': 0.00042562756163788673, 'samples': 2707200, 'steps': 14099, 'loss/train': 1.3508569300174713} +01/27/2022 09:07:31 - INFO - codeparrot_training - Step 14100: {'lr': 0.00042561591652921294, 'samples': 2707392, 'steps': 14100, 'loss/train': 0.8024035692214966} +01/27/2022 09:07:34 - INFO - codeparrot_training - Step 14101: {'lr': 0.00042560427066825636, 'samples': 2707584, 'steps': 14101, 'loss/train': 0.6024859249591827} +01/27/2022 09:07:39 - INFO - codeparrot_training - Step 14102: {'lr': 0.0004255926240550668, 'samples': 2707776, 'steps': 14102, 'loss/train': 1.1285130679607391} +01/27/2022 09:07:42 - INFO - codeparrot_training - Step 14103: {'lr': 0.0004255809766896942, 'samples': 2707968, 'steps': 14103, 'loss/train': 0.7987686395645142} +01/27/2022 09:07:45 - INFO - codeparrot_training - Step 14104: {'lr': 0.00042556932857218855, 'samples': 2708160, 'steps': 14104, 'loss/train': 1.0526391863822937} +01/27/2022 09:07:48 - INFO - codeparrot_training - Step 14105: {'lr': 0.0004255576797025995, 'samples': 2708352, 'steps': 14105, 'loss/train': 1.005784124135971} +01/27/2022 09:07:52 - INFO - codeparrot_training - Step 14106: {'lr': 0.0004255460300809772, 'samples': 2708544, 'steps': 14106, 'loss/train': 0.7830086052417755} +01/27/2022 09:07:55 - INFO - codeparrot_training - Step 14107: {'lr': 0.00042553437970737143, 'samples': 2708736, 'steps': 14107, 'loss/train': 0.6049012541770935} +01/27/2022 09:07:58 - INFO - codeparrot_training - Step 14108: {'lr': 0.00042552272858183203, 'samples': 2708928, 'steps': 14108, 'loss/train': 0.6687837839126587} +01/27/2022 09:08:01 - INFO - codeparrot_training - Step 14109: {'lr': 0.0004255110767044091, 'samples': 2709120, 'steps': 14109, 'loss/train': 0.5699407160282135} +01/27/2022 09:08:04 - INFO - codeparrot_training - Step 14110: {'lr': 0.0004254994240751524, 'samples': 2709312, 'steps': 14110, 'loss/train': 0.8779260814189911} +01/27/2022 09:08:09 - INFO - codeparrot_training - Step 14111: {'lr': 0.00042548777069411194, 'samples': 2709504, 'steps': 14111, 'loss/train': 0.8589128851890564} +01/27/2022 09:08:12 - INFO - codeparrot_training - Step 14112: {'lr': 0.0004254761165613375, 'samples': 2709696, 'steps': 14112, 'loss/train': 1.1824904680252075} +01/27/2022 09:08:15 - INFO - codeparrot_training - Step 14113: {'lr': 0.00042546446167687914, 'samples': 2709888, 'steps': 14113, 'loss/train': 0.7250030189752579} +01/27/2022 09:08:18 - INFO - codeparrot_training - Step 14114: {'lr': 0.00042545280604078673, 'samples': 2710080, 'steps': 14114, 'loss/train': 1.024240404367447} +01/27/2022 09:08:21 - INFO - codeparrot_training - Step 14115: {'lr': 0.0004254411496531103, 'samples': 2710272, 'steps': 14115, 'loss/train': 0.7177295833826065} +01/27/2022 09:08:24 - INFO - codeparrot_training - Step 14116: {'lr': 0.0004254294925138996, 'samples': 2710464, 'steps': 14116, 'loss/train': 0.44890156388282776} +01/27/2022 09:08:27 - INFO - codeparrot_training - Step 14117: {'lr': 0.00042541783462320473, 'samples': 2710656, 'steps': 14117, 'loss/train': 0.8932174444198608} +01/27/2022 09:08:31 - INFO - codeparrot_training - Step 14118: {'lr': 0.00042540617598107544, 'samples': 2710848, 'steps': 14118, 'loss/train': 0.7107881605625153} +01/27/2022 09:08:37 - INFO - codeparrot_training - Step 14119: {'lr': 0.00042539451658756195, 'samples': 2711040, 'steps': 14119, 'loss/train': 1.136687457561493} +01/27/2022 09:08:40 - INFO - codeparrot_training - Step 14120: {'lr': 0.000425382856442714, 'samples': 2711232, 'steps': 14120, 'loss/train': 1.086502343416214} +01/27/2022 09:08:43 - INFO - codeparrot_training - Step 14121: {'lr': 0.0004253711955465815, 'samples': 2711424, 'steps': 14121, 'loss/train': 0.8080064356327057} +01/27/2022 09:08:46 - INFO - codeparrot_training - Step 14122: {'lr': 0.00042535953389921454, 'samples': 2711616, 'steps': 14122, 'loss/train': 0.9183163046836853} +01/27/2022 09:08:49 - INFO - codeparrot_training - Step 14123: {'lr': 0.000425347871500663, 'samples': 2711808, 'steps': 14123, 'loss/train': 0.9480222165584564} +01/27/2022 09:08:53 - INFO - codeparrot_training - Step 14124: {'lr': 0.0004253362083509769, 'samples': 2712000, 'steps': 14124, 'loss/train': 0.9792280197143555} +01/27/2022 09:08:56 - INFO - codeparrot_training - Step 14125: {'lr': 0.0004253245444502061, 'samples': 2712192, 'steps': 14125, 'loss/train': 0.6016221642494202} +01/27/2022 09:08:59 - INFO - codeparrot_training - Step 14126: {'lr': 0.00042531287979840065, 'samples': 2712384, 'steps': 14126, 'loss/train': 0.8418521583080292} +01/27/2022 09:09:02 - INFO - codeparrot_training - Step 14127: {'lr': 0.0004253012143956105, 'samples': 2712576, 'steps': 14127, 'loss/train': 0.24277731031179428} +01/27/2022 09:09:07 - INFO - codeparrot_training - Step 14128: {'lr': 0.0004252895482418856, 'samples': 2712768, 'steps': 14128, 'loss/train': 0.8921669125556946} +01/27/2022 09:09:10 - INFO - codeparrot_training - Step 14129: {'lr': 0.00042527788133727595, 'samples': 2712960, 'steps': 14129, 'loss/train': 0.9275230765342712} +01/27/2022 09:09:13 - INFO - codeparrot_training - Step 14130: {'lr': 0.0004252662136818315, 'samples': 2713152, 'steps': 14130, 'loss/train': 0.7922074198722839} +01/27/2022 09:09:16 - INFO - codeparrot_training - Step 14131: {'lr': 0.00042525454527560225, 'samples': 2713344, 'steps': 14131, 'loss/train': 1.1529269814491272} +01/27/2022 09:09:19 - INFO - codeparrot_training - Step 14132: {'lr': 0.0004252428761186382, 'samples': 2713536, 'steps': 14132, 'loss/train': 0.5792557150125504} +01/27/2022 09:09:22 - INFO - codeparrot_training - Step 14133: {'lr': 0.00042523120621098924, 'samples': 2713728, 'steps': 14133, 'loss/train': 1.1692822873592377} +01/27/2022 09:09:25 - INFO - codeparrot_training - Step 14134: {'lr': 0.0004252195355527055, 'samples': 2713920, 'steps': 14134, 'loss/train': 0.6653560996055603} +01/27/2022 09:09:29 - INFO - codeparrot_training - Step 14135: {'lr': 0.0004252078641438369, 'samples': 2714112, 'steps': 14135, 'loss/train': 1.0650108754634857} +01/27/2022 09:09:32 - INFO - codeparrot_training - Step 14136: {'lr': 0.00042519619198443337, 'samples': 2714304, 'steps': 14136, 'loss/train': 0.9998654723167419} +01/27/2022 09:09:37 - INFO - codeparrot_training - Step 14137: {'lr': 0.0004251845190745451, 'samples': 2714496, 'steps': 14137, 'loss/train': 0.8024831414222717} +01/27/2022 09:09:40 - INFO - codeparrot_training - Step 14138: {'lr': 0.00042517284541422195, 'samples': 2714688, 'steps': 14138, 'loss/train': 0.9910060465335846} +01/27/2022 09:09:43 - INFO - codeparrot_training - Step 14139: {'lr': 0.00042516117100351394, 'samples': 2714880, 'steps': 14139, 'loss/train': 0.7594958245754242} +01/27/2022 09:09:46 - INFO - codeparrot_training - Step 14140: {'lr': 0.0004251494958424711, 'samples': 2715072, 'steps': 14140, 'loss/train': 0.7912977933883667} +01/27/2022 09:09:49 - INFO - codeparrot_training - Step 14141: {'lr': 0.0004251378199311434, 'samples': 2715264, 'steps': 14141, 'loss/train': 0.750194638967514} +01/27/2022 09:09:52 - INFO - codeparrot_training - Step 14142: {'lr': 0.0004251261432695809, 'samples': 2715456, 'steps': 14142, 'loss/train': 0.5675356239080429} +01/27/2022 09:09:55 - INFO - codeparrot_training - Step 14143: {'lr': 0.00042511446585783363, 'samples': 2715648, 'steps': 14143, 'loss/train': 1.613798439502716} +01/27/2022 09:09:59 - INFO - codeparrot_training - Step 14144: {'lr': 0.0004251027876959516, 'samples': 2715840, 'steps': 14144, 'loss/train': 0.8720637559890747} +01/27/2022 09:10:02 - INFO - codeparrot_training - Step 14145: {'lr': 0.0004250911087839848, 'samples': 2716032, 'steps': 14145, 'loss/train': 1.0000647604465485} +01/27/2022 09:10:06 - INFO - codeparrot_training - Step 14146: {'lr': 0.0004250794291219833, 'samples': 2716224, 'steps': 14146, 'loss/train': 1.1733674705028534} +01/27/2022 09:10:09 - INFO - codeparrot_training - Step 14147: {'lr': 0.00042506774870999716, 'samples': 2716416, 'steps': 14147, 'loss/train': 0.2114538475871086} +01/27/2022 09:10:12 - INFO - codeparrot_training - Step 14148: {'lr': 0.00042505606754807634, 'samples': 2716608, 'steps': 14148, 'loss/train': 0.7159367501735687} +01/27/2022 09:10:15 - INFO - codeparrot_training - Step 14149: {'lr': 0.00042504438563627093, 'samples': 2716800, 'steps': 14149, 'loss/train': 1.0990912020206451} +01/27/2022 09:10:19 - INFO - codeparrot_training - Step 14150: {'lr': 0.0004250327029746309, 'samples': 2716992, 'steps': 14150, 'loss/train': 1.0736496448516846} +01/27/2022 09:10:22 - INFO - codeparrot_training - Step 14151: {'lr': 0.0004250210195632064, 'samples': 2717184, 'steps': 14151, 'loss/train': 1.2296795547008514} +01/27/2022 09:10:25 - INFO - codeparrot_training - Step 14152: {'lr': 0.00042500933540204745, 'samples': 2717376, 'steps': 14152, 'loss/train': 0.9675138294696808} +01/27/2022 09:10:28 - INFO - codeparrot_training - Step 14153: {'lr': 0.00042499765049120396, 'samples': 2717568, 'steps': 14153, 'loss/train': 0.4593135863542557} +01/27/2022 09:10:34 - INFO - codeparrot_training - Step 14154: {'lr': 0.0004249859648307263, 'samples': 2717760, 'steps': 14154, 'loss/train': 1.5326351523399353} +01/27/2022 09:10:37 - INFO - codeparrot_training - Step 14155: {'lr': 0.0004249742784206642, 'samples': 2717952, 'steps': 14155, 'loss/train': 0.7025694251060486} +01/27/2022 09:10:41 - INFO - codeparrot_training - Step 14156: {'lr': 0.00042496259126106786, 'samples': 2718144, 'steps': 14156, 'loss/train': 1.1694866716861725} +01/27/2022 09:10:44 - INFO - codeparrot_training - Step 14157: {'lr': 0.00042495090335198735, 'samples': 2718336, 'steps': 14157, 'loss/train': 0.7951121628284454} +01/27/2022 09:10:47 - INFO - codeparrot_training - Step 14158: {'lr': 0.0004249392146934726, 'samples': 2718528, 'steps': 14158, 'loss/train': 0.5545029938220978} +01/27/2022 09:10:50 - INFO - codeparrot_training - Step 14159: {'lr': 0.000424927525285574, 'samples': 2718720, 'steps': 14159, 'loss/train': 0.3984987437725067} +01/27/2022 09:10:53 - INFO - codeparrot_training - Step 14160: {'lr': 0.00042491583512834137, 'samples': 2718912, 'steps': 14160, 'loss/train': 0.5410241782665253} +01/27/2022 09:10:56 - INFO - codeparrot_training - Step 14161: {'lr': 0.00042490414422182484, 'samples': 2719104, 'steps': 14161, 'loss/train': 0.7094369530677795} +01/27/2022 09:11:00 - INFO - codeparrot_training - Step 14162: {'lr': 0.00042489245256607447, 'samples': 2719296, 'steps': 14162, 'loss/train': 1.264877736568451} +01/27/2022 09:11:04 - INFO - codeparrot_training - Step 14163: {'lr': 0.0004248807601611404, 'samples': 2719488, 'steps': 14163, 'loss/train': 1.3792417645454407} +01/27/2022 09:11:07 - INFO - codeparrot_training - Step 14164: {'lr': 0.0004248690670070726, 'samples': 2719680, 'steps': 14164, 'loss/train': 0.8296686708927155} +01/27/2022 09:11:10 - INFO - codeparrot_training - Step 14165: {'lr': 0.00042485737310392135, 'samples': 2719872, 'steps': 14165, 'loss/train': 0.8631308376789093} +01/27/2022 09:11:14 - INFO - codeparrot_training - Step 14166: {'lr': 0.0004248456784517366, 'samples': 2720064, 'steps': 14166, 'loss/train': 0.5005262643098831} +01/27/2022 09:11:17 - INFO - codeparrot_training - Step 14167: {'lr': 0.00042483398305056847, 'samples': 2720256, 'steps': 14167, 'loss/train': 0.13920693844556808} +01/27/2022 09:11:20 - INFO - codeparrot_training - Step 14168: {'lr': 0.0004248222869004671, 'samples': 2720448, 'steps': 14168, 'loss/train': 0.9926440715789795} +01/27/2022 09:11:23 - INFO - codeparrot_training - Step 14169: {'lr': 0.00042481059000148253, 'samples': 2720640, 'steps': 14169, 'loss/train': 0.7302407473325729} +01/27/2022 09:11:26 - INFO - codeparrot_training - Step 14170: {'lr': 0.00042479889235366486, 'samples': 2720832, 'steps': 14170, 'loss/train': 0.7028046548366547} +01/27/2022 09:11:29 - INFO - codeparrot_training - Step 14171: {'lr': 0.0004247871939570643, 'samples': 2721024, 'steps': 14171, 'loss/train': 0.674618661403656} +01/27/2022 09:11:34 - INFO - codeparrot_training - Step 14172: {'lr': 0.00042477549481173093, 'samples': 2721216, 'steps': 14172, 'loss/train': 1.0695421993732452} +01/27/2022 09:11:37 - INFO - codeparrot_training - Step 14173: {'lr': 0.00042476379491771475, 'samples': 2721408, 'steps': 14173, 'loss/train': 0.9074954688549042} +01/27/2022 09:11:40 - INFO - codeparrot_training - Step 14174: {'lr': 0.00042475209427506614, 'samples': 2721600, 'steps': 14174, 'loss/train': 0.5739330053329468} +01/27/2022 09:11:43 - INFO - codeparrot_training - Step 14175: {'lr': 0.00042474039288383484, 'samples': 2721792, 'steps': 14175, 'loss/train': 0.584278866648674} +01/27/2022 09:11:46 - INFO - codeparrot_training - Step 14176: {'lr': 0.0004247286907440713, 'samples': 2721984, 'steps': 14176, 'loss/train': 0.5511540323495865} +01/27/2022 09:11:49 - INFO - codeparrot_training - Step 14177: {'lr': 0.00042471698785582546, 'samples': 2722176, 'steps': 14177, 'loss/train': 0.36209555715322495} +01/27/2022 09:11:53 - INFO - codeparrot_training - Step 14178: {'lr': 0.00042470528421914767, 'samples': 2722368, 'steps': 14178, 'loss/train': 1.1904435753822327} +01/27/2022 09:11:56 - INFO - codeparrot_training - Step 14179: {'lr': 0.0004246935798340877, 'samples': 2722560, 'steps': 14179, 'loss/train': 0.6325263977050781} +01/27/2022 09:12:02 - INFO - codeparrot_training - Step 14180: {'lr': 0.0004246818747006961, 'samples': 2722752, 'steps': 14180, 'loss/train': 0.6864745616912842} +01/27/2022 09:12:05 - INFO - codeparrot_training - Step 14181: {'lr': 0.0004246701688190227, 'samples': 2722944, 'steps': 14181, 'loss/train': 0.8472280204296112} +01/27/2022 09:12:08 - INFO - codeparrot_training - Step 14182: {'lr': 0.0004246584621891179, 'samples': 2723136, 'steps': 14182, 'loss/train': 0.043760388158261776} +01/27/2022 09:12:11 - INFO - codeparrot_training - Step 14183: {'lr': 0.00042464675481103154, 'samples': 2723328, 'steps': 14183, 'loss/train': 0.901787281036377} +01/27/2022 09:12:15 - INFO - codeparrot_training - Step 14184: {'lr': 0.00042463504668481403, 'samples': 2723520, 'steps': 14184, 'loss/train': 1.1302764415740967} +01/27/2022 09:12:18 - INFO - codeparrot_training - Step 14185: {'lr': 0.00042462333781051535, 'samples': 2723712, 'steps': 14185, 'loss/train': 0.8666879832744598} +01/27/2022 09:12:21 - INFO - codeparrot_training - Step 14186: {'lr': 0.00042461162818818585, 'samples': 2723904, 'steps': 14186, 'loss/train': 0.9073822796344757} +01/27/2022 09:12:24 - INFO - codeparrot_training - Step 14187: {'lr': 0.0004245999178178755, 'samples': 2724096, 'steps': 14187, 'loss/train': 0.9063083231449127} +01/27/2022 09:12:27 - INFO - codeparrot_training - Step 14188: {'lr': 0.0004245882066996346, 'samples': 2724288, 'steps': 14188, 'loss/train': 0.935711145401001} +01/27/2022 09:12:32 - INFO - codeparrot_training - Step 14189: {'lr': 0.0004245764948335132, 'samples': 2724480, 'steps': 14189, 'loss/train': 0.1890510767698288} +01/27/2022 09:12:35 - INFO - codeparrot_training - Step 14190: {'lr': 0.0004245647822195616, 'samples': 2724672, 'steps': 14190, 'loss/train': 0.6397561132907867} +01/27/2022 09:12:38 - INFO - codeparrot_training - Step 14191: {'lr': 0.00042455306885782985, 'samples': 2724864, 'steps': 14191, 'loss/train': 0.782521516084671} +01/27/2022 09:12:41 - INFO - codeparrot_training - Step 14192: {'lr': 0.00042454135474836817, 'samples': 2725056, 'steps': 14192, 'loss/train': 1.0616942346096039} +01/27/2022 09:12:44 - INFO - codeparrot_training - Step 14193: {'lr': 0.00042452963989122685, 'samples': 2725248, 'steps': 14193, 'loss/train': 0.6567959189414978} +01/27/2022 09:12:47 - INFO - codeparrot_training - Step 14194: {'lr': 0.00042451792428645587, 'samples': 2725440, 'steps': 14194, 'loss/train': 0.9930361211299896} +01/27/2022 09:12:50 - INFO - codeparrot_training - Step 14195: {'lr': 0.0004245062079341055, 'samples': 2725632, 'steps': 14195, 'loss/train': 0.9698990285396576} +01/27/2022 09:12:54 - INFO - codeparrot_training - Step 14196: {'lr': 0.000424494490834226, 'samples': 2725824, 'steps': 14196, 'loss/train': 0.8933344781398773} +01/27/2022 09:12:57 - INFO - codeparrot_training - Step 14197: {'lr': 0.0004244827729868675, 'samples': 2726016, 'steps': 14197, 'loss/train': 0.6686280816793442} +01/27/2022 09:13:03 - INFO - codeparrot_training - Step 14198: {'lr': 0.00042447105439208024, 'samples': 2726208, 'steps': 14198, 'loss/train': 1.0991587936878204} +01/27/2022 09:13:06 - INFO - codeparrot_training - Step 14199: {'lr': 0.0004244593350499143, 'samples': 2726400, 'steps': 14199, 'loss/train': 0.5852805376052856} +01/27/2022 09:13:09 - INFO - codeparrot_training - Step 14200: {'lr': 0.00042444761496042004, 'samples': 2726592, 'steps': 14200, 'loss/train': 0.8918935060501099} +01/27/2022 09:13:12 - INFO - codeparrot_training - Step 14201: {'lr': 0.0004244358941236476, 'samples': 2726784, 'steps': 14201, 'loss/train': 0.8540935814380646} +01/27/2022 09:13:15 - INFO - codeparrot_training - Step 14202: {'lr': 0.00042442417253964713, 'samples': 2726976, 'steps': 14202, 'loss/train': 0.823919266462326} +01/27/2022 09:13:19 - INFO - codeparrot_training - Step 14203: {'lr': 0.00042441245020846885, 'samples': 2727168, 'steps': 14203, 'loss/train': 0.7311174273490906} +01/27/2022 09:13:22 - INFO - codeparrot_training - Step 14204: {'lr': 0.00042440072713016317, 'samples': 2727360, 'steps': 14204, 'loss/train': 0.901512086391449} +01/27/2022 09:13:25 - INFO - codeparrot_training - Step 14205: {'lr': 0.00042438900330478, 'samples': 2727552, 'steps': 14205, 'loss/train': 1.0390595197677612} +01/27/2022 09:13:28 - INFO - codeparrot_training - Step 14206: {'lr': 0.00042437727873236974, 'samples': 2727744, 'steps': 14206, 'loss/train': 0.6871996968984604} +01/27/2022 09:13:32 - INFO - codeparrot_training - Step 14207: {'lr': 0.00042436555341298266, 'samples': 2727936, 'steps': 14207, 'loss/train': 0.5297298431396484} +01/27/2022 09:13:36 - INFO - codeparrot_training - Step 14208: {'lr': 0.0004243538273466689, 'samples': 2728128, 'steps': 14208, 'loss/train': 0.7079792618751526} +01/27/2022 09:13:39 - INFO - codeparrot_training - Step 14209: {'lr': 0.00042434210053347865, 'samples': 2728320, 'steps': 14209, 'loss/train': 0.5485041439533234} +01/27/2022 09:13:42 - INFO - codeparrot_training - Step 14210: {'lr': 0.0004243303729734622, 'samples': 2728512, 'steps': 14210, 'loss/train': 0.3788662701845169} +01/27/2022 09:13:45 - INFO - codeparrot_training - Step 14211: {'lr': 0.0004243186446666699, 'samples': 2728704, 'steps': 14211, 'loss/train': 0.6294958144426346} +01/27/2022 09:13:48 - INFO - codeparrot_training - Step 14212: {'lr': 0.00042430691561315176, 'samples': 2728896, 'steps': 14212, 'loss/train': 0.659247949719429} +01/27/2022 09:13:51 - INFO - codeparrot_training - Step 14213: {'lr': 0.0004242951858129582, 'samples': 2729088, 'steps': 14213, 'loss/train': 0.53548564016819} +01/27/2022 09:13:55 - INFO - codeparrot_training - Step 14214: {'lr': 0.0004242834552661394, 'samples': 2729280, 'steps': 14214, 'loss/train': 1.1008204221725464} +01/27/2022 09:13:58 - INFO - codeparrot_training - Step 14215: {'lr': 0.0004242717239727456, 'samples': 2729472, 'steps': 14215, 'loss/train': 0.7000266909599304} +01/27/2022 09:14:02 - INFO - codeparrot_training - Step 14216: {'lr': 0.00042425999193282713, 'samples': 2729664, 'steps': 14216, 'loss/train': 0.9997349381446838} +01/27/2022 09:14:05 - INFO - codeparrot_training - Step 14217: {'lr': 0.0004242482591464342, 'samples': 2729856, 'steps': 14217, 'loss/train': 1.5237582921981812} +01/27/2022 09:14:09 - INFO - codeparrot_training - Step 14218: {'lr': 0.0004242365256136169, 'samples': 2730048, 'steps': 14218, 'loss/train': 1.1358030438423157} +01/27/2022 09:14:12 - INFO - codeparrot_training - Step 14219: {'lr': 0.00042422479133442573, 'samples': 2730240, 'steps': 14219, 'loss/train': 0.7082807868719101} +01/27/2022 09:14:15 - INFO - codeparrot_training - Step 14220: {'lr': 0.00042421305630891093, 'samples': 2730432, 'steps': 14220, 'loss/train': 0.5639189332723618} +01/27/2022 09:14:18 - INFO - codeparrot_training - Step 14221: {'lr': 0.0004242013205371227, 'samples': 2730624, 'steps': 14221, 'loss/train': 1.2490195333957672} +01/27/2022 09:14:21 - INFO - codeparrot_training - Step 14222: {'lr': 0.00042418958401911134, 'samples': 2730816, 'steps': 14222, 'loss/train': 0.9368443787097931} +01/27/2022 09:14:24 - INFO - codeparrot_training - Step 14223: {'lr': 0.000424177846754927, 'samples': 2731008, 'steps': 14223, 'loss/train': 0.3169071525335312} +01/27/2022 09:14:29 - INFO - codeparrot_training - Step 14224: {'lr': 0.0004241661087446202, 'samples': 2731200, 'steps': 14224, 'loss/train': 1.0905480980873108} +01/27/2022 09:14:32 - INFO - codeparrot_training - Step 14225: {'lr': 0.00042415436998824105, 'samples': 2731392, 'steps': 14225, 'loss/train': 0.7522726356983185} +01/27/2022 09:14:35 - INFO - codeparrot_training - Step 14226: {'lr': 0.0004241426304858399, 'samples': 2731584, 'steps': 14226, 'loss/train': 0.9421751797199249} +01/27/2022 09:14:38 - INFO - codeparrot_training - Step 14227: {'lr': 0.00042413089023746696, 'samples': 2731776, 'steps': 14227, 'loss/train': 0.2881450653076172} +01/27/2022 09:14:41 - INFO - codeparrot_training - Step 14228: {'lr': 0.00042411914924317265, 'samples': 2731968, 'steps': 14228, 'loss/train': 1.2160905003547668} +01/27/2022 09:14:45 - INFO - codeparrot_training - Step 14229: {'lr': 0.00042410740750300715, 'samples': 2732160, 'steps': 14229, 'loss/train': 1.1507703959941864} +01/27/2022 09:14:48 - INFO - codeparrot_training - Step 14230: {'lr': 0.0004240956650170208, 'samples': 2732352, 'steps': 14230, 'loss/train': 1.1108221113681793} +01/27/2022 09:14:51 - INFO - codeparrot_training - Step 14231: {'lr': 0.00042408392178526396, 'samples': 2732544, 'steps': 14231, 'loss/train': 0.5989001244306564} +01/27/2022 09:14:54 - INFO - codeparrot_training - Step 14232: {'lr': 0.0004240721778077868, 'samples': 2732736, 'steps': 14232, 'loss/train': 0.5962806791067123} +01/27/2022 09:15:00 - INFO - codeparrot_training - Step 14233: {'lr': 0.0004240604330846397, 'samples': 2732928, 'steps': 14233, 'loss/train': 0.8390455842018127} +01/27/2022 09:15:04 - INFO - codeparrot_training - Step 14234: {'lr': 0.000424048687615873, 'samples': 2733120, 'steps': 14234, 'loss/train': 1.0985259711742401} +01/27/2022 09:15:07 - INFO - codeparrot_training - Step 14235: {'lr': 0.00042403694140153705, 'samples': 2733312, 'steps': 14235, 'loss/train': 1.0540865063667297} +01/27/2022 09:15:10 - INFO - codeparrot_training - Step 14236: {'lr': 0.00042402519444168207, 'samples': 2733504, 'steps': 14236, 'loss/train': 0.9751394391059875} +01/27/2022 09:15:13 - INFO - codeparrot_training - Step 14237: {'lr': 0.00042401344673635846, 'samples': 2733696, 'steps': 14237, 'loss/train': 1.0729214251041412} +01/27/2022 09:15:16 - INFO - codeparrot_training - Step 14238: {'lr': 0.00042400169828561636, 'samples': 2733888, 'steps': 14238, 'loss/train': 0.7427747994661331} +01/27/2022 09:15:19 - INFO - codeparrot_training - Step 14239: {'lr': 0.0004239899490895063, 'samples': 2734080, 'steps': 14239, 'loss/train': 0.7567748129367828} +01/27/2022 09:15:23 - INFO - codeparrot_training - Step 14240: {'lr': 0.00042397819914807855, 'samples': 2734272, 'steps': 14240, 'loss/train': 0.9293297231197357} +01/27/2022 09:15:26 - INFO - codeparrot_training - Step 14241: {'lr': 0.00042396644846138355, 'samples': 2734464, 'steps': 14241, 'loss/train': 0.9282006025314331} +01/27/2022 09:15:30 - INFO - codeparrot_training - Step 14242: {'lr': 0.00042395469702947135, 'samples': 2734656, 'steps': 14242, 'loss/train': 0.26918093115091324} +01/27/2022 09:15:33 - INFO - codeparrot_training - Step 14243: {'lr': 0.0004239429448523925, 'samples': 2734848, 'steps': 14243, 'loss/train': 0.544265404343605} +01/27/2022 09:15:37 - INFO - codeparrot_training - Step 14244: {'lr': 0.00042393119193019743, 'samples': 2735040, 'steps': 14244, 'loss/train': 0.8511837422847748} +01/27/2022 09:15:40 - INFO - codeparrot_training - Step 14245: {'lr': 0.00042391943826293623, 'samples': 2735232, 'steps': 14245, 'loss/train': 0.6432247906923294} +01/27/2022 09:15:43 - INFO - codeparrot_training - Step 14246: {'lr': 0.0004239076838506595, 'samples': 2735424, 'steps': 14246, 'loss/train': 0.7615760564804077} +01/27/2022 09:15:46 - INFO - codeparrot_training - Step 14247: {'lr': 0.0004238959286934174, 'samples': 2735616, 'steps': 14247, 'loss/train': 0.9416446387767792} +01/27/2022 09:15:49 - INFO - codeparrot_training - Step 14248: {'lr': 0.0004238841727912603, 'samples': 2735808, 'steps': 14248, 'loss/train': 0.9987920522689819} +01/27/2022 09:15:52 - INFO - codeparrot_training - Step 14249: {'lr': 0.00042387241614423875, 'samples': 2736000, 'steps': 14249, 'loss/train': 0.8325115442276001} +01/27/2022 09:15:55 - INFO - codeparrot_training - Step 14250: {'lr': 0.0004238606587524029, 'samples': 2736192, 'steps': 14250, 'loss/train': 0.937267005443573} +01/27/2022 09:16:00 - INFO - codeparrot_training - Step 14251: {'lr': 0.0004238489006158033, 'samples': 2736384, 'steps': 14251, 'loss/train': 1.268635869026184} +01/27/2022 09:16:03 - INFO - codeparrot_training - Step 14252: {'lr': 0.00042383714173449007, 'samples': 2736576, 'steps': 14252, 'loss/train': 1.4034784734249115} +01/27/2022 09:16:06 - INFO - codeparrot_training - Step 14253: {'lr': 0.0004238253821085138, 'samples': 2736768, 'steps': 14253, 'loss/train': 1.039599359035492} +01/27/2022 09:16:09 - INFO - codeparrot_training - Step 14254: {'lr': 0.00042381362173792475, 'samples': 2736960, 'steps': 14254, 'loss/train': 1.2722280621528625} +01/27/2022 09:16:13 - INFO - codeparrot_training - Step 14255: {'lr': 0.00042380186062277337, 'samples': 2737152, 'steps': 14255, 'loss/train': 0.7872245907783508} +01/27/2022 09:16:16 - INFO - codeparrot_training - Step 14256: {'lr': 0.00042379009876311, 'samples': 2737344, 'steps': 14256, 'loss/train': 0.624341681599617} +01/27/2022 09:16:19 - INFO - codeparrot_training - Step 14257: {'lr': 0.00042377833615898496, 'samples': 2737536, 'steps': 14257, 'loss/train': 0.6198979765176773} +01/27/2022 09:16:22 - INFO - codeparrot_training - Step 14258: {'lr': 0.0004237665728104488, 'samples': 2737728, 'steps': 14258, 'loss/train': 0.6413855999708176} +01/27/2022 09:16:25 - INFO - codeparrot_training - Step 14259: {'lr': 0.0004237548087175518, 'samples': 2737920, 'steps': 14259, 'loss/train': 0.36263226717710495} +01/27/2022 09:16:31 - INFO - codeparrot_training - Step 14260: {'lr': 0.00042374304388034437, 'samples': 2738112, 'steps': 14260, 'loss/train': 0.9732244312763214} +01/27/2022 09:16:34 - INFO - codeparrot_training - Step 14261: {'lr': 0.00042373127829887694, 'samples': 2738304, 'steps': 14261, 'loss/train': 0.7516420483589172} +01/27/2022 09:16:38 - INFO - codeparrot_training - Step 14262: {'lr': 0.0004237195119731998, 'samples': 2738496, 'steps': 14262, 'loss/train': 0.5315220057964325} +01/27/2022 09:16:41 - INFO - codeparrot_training - Step 14263: {'lr': 0.0004237077449033635, 'samples': 2738688, 'steps': 14263, 'loss/train': 0.7832159399986267} +01/27/2022 09:16:44 - INFO - codeparrot_training - Step 14264: {'lr': 0.0004236959770894183, 'samples': 2738880, 'steps': 14264, 'loss/train': 0.9474566280841827} +01/27/2022 09:16:47 - INFO - codeparrot_training - Step 14265: {'lr': 0.0004236842085314148, 'samples': 2739072, 'steps': 14265, 'loss/train': 0.586104691028595} +01/27/2022 09:16:50 - INFO - codeparrot_training - Step 14266: {'lr': 0.0004236724392294032, 'samples': 2739264, 'steps': 14266, 'loss/train': 1.1891035437583923} +01/27/2022 09:16:53 - INFO - codeparrot_training - Step 14267: {'lr': 0.0004236606691834341, 'samples': 2739456, 'steps': 14267, 'loss/train': 0.666012167930603} +01/27/2022 09:16:58 - INFO - codeparrot_training - Step 14268: {'lr': 0.0004236488983935578, 'samples': 2739648, 'steps': 14268, 'loss/train': 0.8797098398208618} +01/27/2022 09:17:01 - INFO - codeparrot_training - Step 14269: {'lr': 0.0004236371268598248, 'samples': 2739840, 'steps': 14269, 'loss/train': 0.6376208961009979} +01/27/2022 09:17:04 - INFO - codeparrot_training - Step 14270: {'lr': 0.0004236253545822855, 'samples': 2740032, 'steps': 14270, 'loss/train': 0.5103703290224075} +01/27/2022 09:17:07 - INFO - codeparrot_training - Step 14271: {'lr': 0.00042361358156099016, 'samples': 2740224, 'steps': 14271, 'loss/train': 0.8884707391262054} +01/27/2022 09:17:10 - INFO - codeparrot_training - Step 14272: {'lr': 0.0004236018077959895, 'samples': 2740416, 'steps': 14272, 'loss/train': 0.20819831639528275} +01/27/2022 09:17:14 - INFO - codeparrot_training - Step 14273: {'lr': 0.00042359003328733383, 'samples': 2740608, 'steps': 14273, 'loss/train': 0.9746081829071045} +01/27/2022 09:17:17 - INFO - codeparrot_training - Step 14274: {'lr': 0.0004235782580350734, 'samples': 2740800, 'steps': 14274, 'loss/train': 0.5404078960418701} +01/27/2022 09:17:20 - INFO - codeparrot_training - Step 14275: {'lr': 0.0004235664820392591, 'samples': 2740992, 'steps': 14275, 'loss/train': 0.6333228796720505} +01/27/2022 09:17:23 - INFO - codeparrot_training - Step 14276: {'lr': 0.0004235547052999409, 'samples': 2741184, 'steps': 14276, 'loss/train': 1.0300936102867126} +01/27/2022 09:17:29 - INFO - codeparrot_training - Step 14277: {'lr': 0.0004235429278171695, 'samples': 2741376, 'steps': 14277, 'loss/train': 1.0848921239376068} +01/27/2022 09:17:32 - INFO - codeparrot_training - Step 14278: {'lr': 0.00042353114959099535, 'samples': 2741568, 'steps': 14278, 'loss/train': 0.7284392416477203} +01/27/2022 09:17:36 - INFO - codeparrot_training - Step 14279: {'lr': 0.0004235193706214688, 'samples': 2741760, 'steps': 14279, 'loss/train': 0.3238080069422722} +01/27/2022 09:17:39 - INFO - codeparrot_training - Step 14280: {'lr': 0.00042350759090864043, 'samples': 2741952, 'steps': 14280, 'loss/train': 0.9572981894016266} +01/27/2022 09:17:42 - INFO - codeparrot_training - Step 14281: {'lr': 0.00042349581045256055, 'samples': 2742144, 'steps': 14281, 'loss/train': 0.5750007033348083} +01/27/2022 09:17:45 - INFO - codeparrot_training - Step 14282: {'lr': 0.00042348402925327977, 'samples': 2742336, 'steps': 14282, 'loss/train': 0.8269672393798828} +01/27/2022 09:17:48 - INFO - codeparrot_training - Step 14283: {'lr': 0.00042347224731084854, 'samples': 2742528, 'steps': 14283, 'loss/train': 0.5129463225603104} +01/27/2022 09:17:51 - INFO - codeparrot_training - Step 14284: {'lr': 0.0004234604646253172, 'samples': 2742720, 'steps': 14284, 'loss/train': 1.0970392227172852} +01/27/2022 09:17:54 - INFO - codeparrot_training - Step 14285: {'lr': 0.0004234486811967364, 'samples': 2742912, 'steps': 14285, 'loss/train': 2.0517890453338623} +01/27/2022 09:17:59 - INFO - codeparrot_training - Step 14286: {'lr': 0.00042343689702515643, 'samples': 2743104, 'steps': 14286, 'loss/train': 0.9682847857475281} +01/27/2022 09:18:02 - INFO - codeparrot_training - Step 14287: {'lr': 0.0004234251121106279, 'samples': 2743296, 'steps': 14287, 'loss/train': 1.1051902770996094} +01/27/2022 09:18:05 - INFO - codeparrot_training - Step 14288: {'lr': 0.00042341332645320126, 'samples': 2743488, 'steps': 14288, 'loss/train': 0.5120789408683777} +01/27/2022 09:18:08 - INFO - codeparrot_training - Step 14289: {'lr': 0.000423401540052927, 'samples': 2743680, 'steps': 14289, 'loss/train': 0.5988226979970932} +01/27/2022 09:18:12 - INFO - codeparrot_training - Step 14290: {'lr': 0.0004233897529098556, 'samples': 2743872, 'steps': 14290, 'loss/train': 0.9767035245895386} +01/27/2022 09:18:15 - INFO - codeparrot_training - Step 14291: {'lr': 0.0004233779650240376, 'samples': 2744064, 'steps': 14291, 'loss/train': 1.0035127401351929} +01/27/2022 09:18:18 - INFO - codeparrot_training - Step 14292: {'lr': 0.00042336617639552335, 'samples': 2744256, 'steps': 14292, 'loss/train': 0.7534035444259644} +01/27/2022 09:18:21 - INFO - codeparrot_training - Step 14293: {'lr': 0.00042335438702436354, 'samples': 2744448, 'steps': 14293, 'loss/train': 0.8752596974372864} +01/27/2022 09:18:24 - INFO - codeparrot_training - Step 14294: {'lr': 0.0004233425969106085, 'samples': 2744640, 'steps': 14294, 'loss/train': 1.3263380527496338} +01/27/2022 09:18:29 - INFO - codeparrot_training - Step 14295: {'lr': 0.00042333080605430883, 'samples': 2744832, 'steps': 14295, 'loss/train': 0.7577779591083527} +01/27/2022 09:18:32 - INFO - codeparrot_training - Step 14296: {'lr': 0.00042331901445551514, 'samples': 2745024, 'steps': 14296, 'loss/train': 0.7451578080654144} +01/27/2022 09:18:35 - INFO - codeparrot_training - Step 14297: {'lr': 0.00042330722211427775, 'samples': 2745216, 'steps': 14297, 'loss/train': 0.9192924499511719} +01/27/2022 09:18:38 - INFO - codeparrot_training - Step 14298: {'lr': 0.00042329542903064724, 'samples': 2745408, 'steps': 14298, 'loss/train': 1.1104785203933716} +01/27/2022 09:18:41 - INFO - codeparrot_training - Step 14299: {'lr': 0.00042328363520467417, 'samples': 2745600, 'steps': 14299, 'loss/train': 0.9583881497383118} +01/27/2022 09:18:44 - INFO - codeparrot_training - Step 14300: {'lr': 0.000423271840636409, 'samples': 2745792, 'steps': 14300, 'loss/train': 0.513668954372406} +01/27/2022 09:18:47 - INFO - codeparrot_training - Step 14301: {'lr': 0.0004232600453259023, 'samples': 2745984, 'steps': 14301, 'loss/train': 1.0602406561374664} +01/27/2022 09:18:51 - INFO - codeparrot_training - Step 14302: {'lr': 0.0004232482492732046, 'samples': 2746176, 'steps': 14302, 'loss/train': 0.9703687727451324} +01/27/2022 09:18:54 - INFO - codeparrot_training - Step 14303: {'lr': 0.00042323645247836636, 'samples': 2746368, 'steps': 14303, 'loss/train': 0.4998127520084381} +01/27/2022 09:19:00 - INFO - codeparrot_training - Step 14304: {'lr': 0.00042322465494143814, 'samples': 2746560, 'steps': 14304, 'loss/train': 0.5433225631713867} +01/27/2022 09:19:03 - INFO - codeparrot_training - Step 14305: {'lr': 0.00042321285666247063, 'samples': 2746752, 'steps': 14305, 'loss/train': 0.933937668800354} +01/27/2022 09:19:06 - INFO - codeparrot_training - Step 14306: {'lr': 0.0004232010576415141, 'samples': 2746944, 'steps': 14306, 'loss/train': 0.7979668378829956} +01/27/2022 09:19:09 - INFO - codeparrot_training - Step 14307: {'lr': 0.00042318925787861937, 'samples': 2747136, 'steps': 14307, 'loss/train': 0.5648932009935379} +01/27/2022 09:19:12 - INFO - codeparrot_training - Step 14308: {'lr': 0.0004231774573738367, 'samples': 2747328, 'steps': 14308, 'loss/train': 0.5788000077009201} +01/27/2022 09:19:16 - INFO - codeparrot_training - Step 14309: {'lr': 0.000423165656127217, 'samples': 2747520, 'steps': 14309, 'loss/train': 0.37852974236011505} +01/27/2022 09:19:19 - INFO - codeparrot_training - Step 14310: {'lr': 0.00042315385413881047, 'samples': 2747712, 'steps': 14310, 'loss/train': 1.3541596233844757} +01/27/2022 09:19:22 - INFO - codeparrot_training - Step 14311: {'lr': 0.00042314205140866785, 'samples': 2747904, 'steps': 14311, 'loss/train': 0.9080301225185394} +01/27/2022 09:19:27 - INFO - codeparrot_training - Step 14312: {'lr': 0.00042313024793683965, 'samples': 2748096, 'steps': 14312, 'loss/train': 0.6868321895599365} +01/27/2022 09:19:30 - INFO - codeparrot_training - Step 14313: {'lr': 0.0004231184437233765, 'samples': 2748288, 'steps': 14313, 'loss/train': 0.7418138980865479} +01/27/2022 09:19:33 - INFO - codeparrot_training - Step 14314: {'lr': 0.0004231066387683288, 'samples': 2748480, 'steps': 14314, 'loss/train': 0.8356214761734009} +01/27/2022 09:19:36 - INFO - codeparrot_training - Step 14315: {'lr': 0.0004230948330717472, 'samples': 2748672, 'steps': 14315, 'loss/train': 0.618102103471756} +01/27/2022 09:19:40 - INFO - codeparrot_training - Step 14316: {'lr': 0.0004230830266336825, 'samples': 2748864, 'steps': 14316, 'loss/train': 0.586384579539299} +01/27/2022 09:19:43 - INFO - codeparrot_training - Step 14317: {'lr': 0.00042307121945418493, 'samples': 2749056, 'steps': 14317, 'loss/train': 0.4852640926837921} +01/27/2022 09:19:46 - INFO - codeparrot_training - Step 14318: {'lr': 0.00042305941153330525, 'samples': 2749248, 'steps': 14318, 'loss/train': 1.0242800116539001} +01/27/2022 09:19:49 - INFO - codeparrot_training - Step 14319: {'lr': 0.00042304760287109394, 'samples': 2749440, 'steps': 14319, 'loss/train': 0.8425469398498535} +01/27/2022 09:19:52 - INFO - codeparrot_training - Step 14320: {'lr': 0.0004230357934676017, 'samples': 2749632, 'steps': 14320, 'loss/train': 0.4784736782312393} +01/27/2022 09:19:55 - INFO - codeparrot_training - Step 14321: {'lr': 0.00042302398332287903, 'samples': 2749824, 'steps': 14321, 'loss/train': 0.251325823366642} +01/27/2022 09:20:00 - INFO - codeparrot_training - Step 14322: {'lr': 0.00042301217243697665, 'samples': 2750016, 'steps': 14322, 'loss/train': 0.8205897510051727} +01/27/2022 09:20:03 - INFO - codeparrot_training - Step 14323: {'lr': 0.00042300036080994495, 'samples': 2750208, 'steps': 14323, 'loss/train': 0.6214554458856583} +01/27/2022 09:20:06 - INFO - codeparrot_training - Step 14324: {'lr': 0.00042298854844183476, 'samples': 2750400, 'steps': 14324, 'loss/train': 0.7097420990467072} +01/27/2022 09:20:09 - INFO - codeparrot_training - Step 14325: {'lr': 0.0004229767353326964, 'samples': 2750592, 'steps': 14325, 'loss/train': 0.3682222366333008} +01/27/2022 09:20:12 - INFO - codeparrot_training - Step 14326: {'lr': 0.0004229649214825808, 'samples': 2750784, 'steps': 14326, 'loss/train': 0.588453009724617} +01/27/2022 09:20:16 - INFO - codeparrot_training - Step 14327: {'lr': 0.0004229531068915383, 'samples': 2750976, 'steps': 14327, 'loss/train': 1.2690320312976837} +01/27/2022 09:20:19 - INFO - codeparrot_training - Step 14328: {'lr': 0.0004229412915596196, 'samples': 2751168, 'steps': 14328, 'loss/train': 1.0698515474796295} +01/27/2022 09:20:22 - INFO - codeparrot_training - Step 14329: {'lr': 0.0004229294754868754, 'samples': 2751360, 'steps': 14329, 'loss/train': 0.6015733927488327} +01/27/2022 09:20:25 - INFO - codeparrot_training - Step 14330: {'lr': 0.0004229176586733562, 'samples': 2751552, 'steps': 14330, 'loss/train': 0.6864897608757019} +01/27/2022 09:20:29 - INFO - codeparrot_training - Step 14331: {'lr': 0.0004229058411191126, 'samples': 2751744, 'steps': 14331, 'loss/train': 0.7860829532146454} +01/27/2022 09:20:32 - INFO - codeparrot_training - Step 14332: {'lr': 0.0004228940228241953, 'samples': 2751936, 'steps': 14332, 'loss/train': 0.8003464937210083} +01/27/2022 09:20:36 - INFO - codeparrot_training - Step 14333: {'lr': 0.0004228822037886549, 'samples': 2752128, 'steps': 14333, 'loss/train': 0.5524534732103348} +01/27/2022 09:20:39 - INFO - codeparrot_training - Step 14334: {'lr': 0.00042287038401254214, 'samples': 2752320, 'steps': 14334, 'loss/train': 0.94110506772995} +01/27/2022 09:20:42 - INFO - codeparrot_training - Step 14335: {'lr': 0.00042285856349590746, 'samples': 2752512, 'steps': 14335, 'loss/train': 1.186696618795395} +01/27/2022 09:20:45 - INFO - codeparrot_training - Step 14336: {'lr': 0.0004228467422388016, 'samples': 2752704, 'steps': 14336, 'loss/train': 0.5758137702941895} +01/27/2022 09:20:48 - INFO - codeparrot_training - Step 14337: {'lr': 0.00042283492024127524, 'samples': 2752896, 'steps': 14337, 'loss/train': 1.2521802484989166} +01/27/2022 09:20:51 - INFO - codeparrot_training - Step 14338: {'lr': 0.00042282309750337887, 'samples': 2753088, 'steps': 14338, 'loss/train': 0.534479945898056} +01/27/2022 09:20:57 - INFO - codeparrot_training - Step 14339: {'lr': 0.0004228112740251632, 'samples': 2753280, 'steps': 14339, 'loss/train': 0.7077648192644119} +01/27/2022 09:21:01 - INFO - codeparrot_training - Step 14340: {'lr': 0.00042279944980667906, 'samples': 2753472, 'steps': 14340, 'loss/train': 0.7233779579401016} +01/27/2022 09:21:04 - INFO - codeparrot_training - Step 14341: {'lr': 0.00042278762484797684, 'samples': 2753664, 'steps': 14341, 'loss/train': 0.43581072986125946} +01/27/2022 09:21:07 - INFO - codeparrot_training - Step 14342: {'lr': 0.0004227757991491073, 'samples': 2753856, 'steps': 14342, 'loss/train': 1.0082566738128662} +01/27/2022 09:21:10 - INFO - codeparrot_training - Step 14343: {'lr': 0.0004227639727101211, 'samples': 2754048, 'steps': 14343, 'loss/train': 1.0097726583480835} +01/27/2022 09:21:13 - INFO - codeparrot_training - Step 14344: {'lr': 0.0004227521455310689, 'samples': 2754240, 'steps': 14344, 'loss/train': 0.38958849012851715} +01/27/2022 09:21:16 - INFO - codeparrot_training - Step 14345: {'lr': 0.0004227403176120014, 'samples': 2754432, 'steps': 14345, 'loss/train': 1.1212505400180817} +01/27/2022 09:21:19 - INFO - codeparrot_training - Step 14346: {'lr': 0.00042272848895296924, 'samples': 2754624, 'steps': 14346, 'loss/train': 0.6210549473762512} +01/27/2022 09:21:23 - INFO - codeparrot_training - Step 14347: {'lr': 0.000422716659554023, 'samples': 2754816, 'steps': 14347, 'loss/train': 1.1838494539260864} +01/27/2022 09:21:27 - INFO - codeparrot_training - Step 14348: {'lr': 0.00042270482941521347, 'samples': 2755008, 'steps': 14348, 'loss/train': 0.5793857127428055} +01/27/2022 09:21:30 - INFO - codeparrot_training - Step 14349: {'lr': 0.0004226929985365913, 'samples': 2755200, 'steps': 14349, 'loss/train': 1.1945917010307312} +01/27/2022 09:21:33 - INFO - codeparrot_training - Step 14350: {'lr': 0.00042268116691820723, 'samples': 2755392, 'steps': 14350, 'loss/train': 0.22134210169315338} +01/27/2022 09:21:36 - INFO - codeparrot_training - Step 14351: {'lr': 0.00042266933456011174, 'samples': 2755584, 'steps': 14351, 'loss/train': 0.5421672910451889} +01/27/2022 09:21:40 - INFO - codeparrot_training - Step 14352: {'lr': 0.0004226575014623557, 'samples': 2755776, 'steps': 14352, 'loss/train': 0.7268961668014526} +01/27/2022 09:21:43 - INFO - codeparrot_training - Step 14353: {'lr': 0.0004226456676249898, 'samples': 2755968, 'steps': 14353, 'loss/train': 1.1724359393119812} +01/27/2022 09:21:46 - INFO - codeparrot_training - Step 14354: {'lr': 0.0004226338330480646, 'samples': 2756160, 'steps': 14354, 'loss/train': 0.5814786404371262} +01/27/2022 09:21:49 - INFO - codeparrot_training - Step 14355: {'lr': 0.00042262199773163096, 'samples': 2756352, 'steps': 14355, 'loss/train': 1.291445016860962} +01/27/2022 09:21:52 - INFO - codeparrot_training - Step 14356: {'lr': 0.00042261016167573944, 'samples': 2756544, 'steps': 14356, 'loss/train': 0.8525928854942322} +01/27/2022 09:21:58 - INFO - codeparrot_training - Step 14357: {'lr': 0.0004225983248804408, 'samples': 2756736, 'steps': 14357, 'loss/train': 0.9238451421260834} +01/27/2022 09:22:01 - INFO - codeparrot_training - Step 14358: {'lr': 0.0004225864873457858, 'samples': 2756928, 'steps': 14358, 'loss/train': 0.5528281331062317} +01/27/2022 09:22:04 - INFO - codeparrot_training - Step 14359: {'lr': 0.0004225746490718251, 'samples': 2757120, 'steps': 14359, 'loss/train': 0.8927678167819977} +01/27/2022 09:22:08 - INFO - codeparrot_training - Step 14360: {'lr': 0.0004225628100586093, 'samples': 2757312, 'steps': 14360, 'loss/train': 0.8085581660270691} +01/27/2022 09:22:11 - INFO - codeparrot_training - Step 14361: {'lr': 0.0004225509703061893, 'samples': 2757504, 'steps': 14361, 'loss/train': 0.7228201925754547} +01/27/2022 09:22:14 - INFO - codeparrot_training - Step 14362: {'lr': 0.0004225391298146157, 'samples': 2757696, 'steps': 14362, 'loss/train': 0.44779010117053986} +01/27/2022 09:22:17 - INFO - codeparrot_training - Step 14363: {'lr': 0.0004225272885839392, 'samples': 2757888, 'steps': 14363, 'loss/train': 0.8764284253120422} +01/27/2022 09:22:20 - INFO - codeparrot_training - Step 14364: {'lr': 0.0004225154466142107, 'samples': 2758080, 'steps': 14364, 'loss/train': 1.0763943493366241} +01/27/2022 09:22:24 - INFO - codeparrot_training - Step 14365: {'lr': 0.0004225036039054807, 'samples': 2758272, 'steps': 14365, 'loss/train': 0.45451095700263977} +01/27/2022 09:22:28 - INFO - codeparrot_training - Step 14366: {'lr': 0.00042249176045780013, 'samples': 2758464, 'steps': 14366, 'loss/train': 2.0474377870559692} +01/27/2022 09:22:31 - INFO - codeparrot_training - Step 14367: {'lr': 0.0004224799162712195, 'samples': 2758656, 'steps': 14367, 'loss/train': 0.9160047769546509} +01/27/2022 09:22:34 - INFO - codeparrot_training - Step 14368: {'lr': 0.0004224680713457898, 'samples': 2758848, 'steps': 14368, 'loss/train': 0.6571607440710068} +01/27/2022 09:22:37 - INFO - codeparrot_training - Step 14369: {'lr': 0.00042245622568156164, 'samples': 2759040, 'steps': 14369, 'loss/train': 0.8708453178405762} +01/27/2022 09:22:40 - INFO - codeparrot_training - Step 14370: {'lr': 0.0004224443792785857, 'samples': 2759232, 'steps': 14370, 'loss/train': 0.5567598938941956} +01/27/2022 09:22:43 - INFO - codeparrot_training - Step 14371: {'lr': 0.0004224325321369128, 'samples': 2759424, 'steps': 14371, 'loss/train': 1.2571393847465515} +01/27/2022 09:22:47 - INFO - codeparrot_training - Step 14372: {'lr': 0.0004224206842565937, 'samples': 2759616, 'steps': 14372, 'loss/train': 0.6731864959001541} +01/27/2022 09:22:50 - INFO - codeparrot_training - Step 14373: {'lr': 0.00042240883563767916, 'samples': 2759808, 'steps': 14373, 'loss/train': 0.8399308919906616} +01/27/2022 09:22:54 - INFO - codeparrot_training - Step 14374: {'lr': 0.00042239698628021994, 'samples': 2760000, 'steps': 14374, 'loss/train': 0.9541171789169312} +01/27/2022 09:22:57 - INFO - codeparrot_training - Step 14375: {'lr': 0.0004223851361842668, 'samples': 2760192, 'steps': 14375, 'loss/train': 1.4306694567203522} +01/27/2022 09:23:00 - INFO - codeparrot_training - Step 14376: {'lr': 0.00042237328534987034, 'samples': 2760384, 'steps': 14376, 'loss/train': 1.4691174924373627} +01/27/2022 09:23:04 - INFO - codeparrot_training - Step 14377: {'lr': 0.0004223614337770816, 'samples': 2760576, 'steps': 14377, 'loss/train': 0.7572878301143646} +01/27/2022 09:23:07 - INFO - codeparrot_training - Step 14378: {'lr': 0.0004223495814659511, 'samples': 2760768, 'steps': 14378, 'loss/train': 0.8255065977573395} +01/27/2022 09:23:10 - INFO - codeparrot_training - Step 14379: {'lr': 0.00042233772841652974, 'samples': 2760960, 'steps': 14379, 'loss/train': 0.836522787809372} +01/27/2022 09:23:13 - INFO - codeparrot_training - Step 14380: {'lr': 0.00042232587462886833, 'samples': 2761152, 'steps': 14380, 'loss/train': 0.7791426479816437} +01/27/2022 09:23:16 - INFO - codeparrot_training - Step 14381: {'lr': 0.0004223140201030176, 'samples': 2761344, 'steps': 14381, 'loss/train': 0.5605128407478333} +01/27/2022 09:23:19 - INFO - codeparrot_training - Step 14382: {'lr': 0.0004223021648390283, 'samples': 2761536, 'steps': 14382, 'loss/train': 0.8333341777324677} +01/27/2022 09:23:26 - INFO - codeparrot_training - Step 14383: {'lr': 0.0004222903088369512, 'samples': 2761728, 'steps': 14383, 'loss/train': 0.14003561809659004} +01/27/2022 09:23:29 - INFO - codeparrot_training - Step 14384: {'lr': 0.0004222784520968371, 'samples': 2761920, 'steps': 14384, 'loss/train': 0.7671217918395996} +01/27/2022 09:23:32 - INFO - codeparrot_training - Step 14385: {'lr': 0.000422266594618737, 'samples': 2762112, 'steps': 14385, 'loss/train': 0.7966287732124329} +01/27/2022 09:23:35 - INFO - codeparrot_training - Step 14386: {'lr': 0.0004222547364027013, 'samples': 2762304, 'steps': 14386, 'loss/train': 1.267805814743042} +01/27/2022 09:23:39 - INFO - codeparrot_training - Step 14387: {'lr': 0.0004222428774487811, 'samples': 2762496, 'steps': 14387, 'loss/train': 1.0170113146305084} +01/27/2022 09:23:42 - INFO - codeparrot_training - Step 14388: {'lr': 0.00042223101775702704, 'samples': 2762688, 'steps': 14388, 'loss/train': 0.8901536464691162} +01/27/2022 09:23:45 - INFO - codeparrot_training - Step 14389: {'lr': 0.00042221915732749006, 'samples': 2762880, 'steps': 14389, 'loss/train': 0.7059946954250336} +01/27/2022 09:23:48 - INFO - codeparrot_training - Step 14390: {'lr': 0.0004222072961602209, 'samples': 2763072, 'steps': 14390, 'loss/train': 1.028389424085617} +01/27/2022 09:23:51 - INFO - codeparrot_training - Step 14391: {'lr': 0.0004221954342552703, 'samples': 2763264, 'steps': 14391, 'loss/train': 0.895744800567627} +01/27/2022 09:23:54 - INFO - codeparrot_training - Step 14392: {'lr': 0.00042218357161268917, 'samples': 2763456, 'steps': 14392, 'loss/train': 0.9488701522350311} +01/27/2022 09:23:59 - INFO - codeparrot_training - Step 14393: {'lr': 0.0004221717082325283, 'samples': 2763648, 'steps': 14393, 'loss/train': 0.6655298620462418} +01/27/2022 09:24:02 - INFO - codeparrot_training - Step 14394: {'lr': 0.00042215984411483854, 'samples': 2763840, 'steps': 14394, 'loss/train': 0.7588776648044586} +01/27/2022 09:24:05 - INFO - codeparrot_training - Step 14395: {'lr': 0.00042214797925967064, 'samples': 2764032, 'steps': 14395, 'loss/train': 1.2940278053283691} +01/27/2022 09:24:08 - INFO - codeparrot_training - Step 14396: {'lr': 0.00042213611366707547, 'samples': 2764224, 'steps': 14396, 'loss/train': 0.8323528468608856} +01/27/2022 09:24:12 - INFO - codeparrot_training - Step 14397: {'lr': 0.0004221242473371038, 'samples': 2764416, 'steps': 14397, 'loss/train': 1.0071116387844086} +01/27/2022 09:24:15 - INFO - codeparrot_training - Step 14398: {'lr': 0.00042211238026980657, 'samples': 2764608, 'steps': 14398, 'loss/train': 0.9704278707504272} +01/27/2022 09:24:18 - INFO - codeparrot_training - Step 14399: {'lr': 0.0004221005124652345, 'samples': 2764800, 'steps': 14399, 'loss/train': 1.021740585565567} +01/27/2022 09:24:21 - INFO - codeparrot_training - Step 14400: {'lr': 0.0004220886439234385, 'samples': 2764992, 'steps': 14400, 'loss/train': 1.0850888192653656} +01/27/2022 09:24:24 - INFO - codeparrot_training - Step 14401: {'lr': 0.0004220767746444694, 'samples': 2765184, 'steps': 14401, 'loss/train': 0.7459539324045181} +01/27/2022 09:24:30 - INFO - codeparrot_training - Step 14402: {'lr': 0.0004220649046283781, 'samples': 2765376, 'steps': 14402, 'loss/train': 0.9197062253952026} +01/27/2022 09:24:33 - INFO - codeparrot_training - Step 14403: {'lr': 0.00042205303387521533, 'samples': 2765568, 'steps': 14403, 'loss/train': 1.0628364980220795} +01/27/2022 09:24:37 - INFO - codeparrot_training - Step 14404: {'lr': 0.00042204116238503197, 'samples': 2765760, 'steps': 14404, 'loss/train': 0.5122514516115189} +01/27/2022 09:24:40 - INFO - codeparrot_training - Step 14405: {'lr': 0.00042202929015787893, 'samples': 2765952, 'steps': 14405, 'loss/train': 0.04109922889620066} +01/27/2022 09:24:43 - INFO - codeparrot_training - Step 14406: {'lr': 0.000422017417193807, 'samples': 2766144, 'steps': 14406, 'loss/train': 0.4987386167049408} +01/27/2022 09:24:46 - INFO - codeparrot_training - Step 14407: {'lr': 0.0004220055434928671, 'samples': 2766336, 'steps': 14407, 'loss/train': 1.1031387448310852} +01/27/2022 09:24:49 - INFO - codeparrot_training - Step 14408: {'lr': 0.0004219936690551101, 'samples': 2766528, 'steps': 14408, 'loss/train': 1.2550359964370728} +01/27/2022 09:24:53 - INFO - codeparrot_training - Step 14409: {'lr': 0.0004219817938805869, 'samples': 2766720, 'steps': 14409, 'loss/train': 0.839638352394104} +01/27/2022 09:24:57 - INFO - codeparrot_training - Step 14410: {'lr': 0.0004219699179693481, 'samples': 2766912, 'steps': 14410, 'loss/train': 0.9101841151714325} +01/27/2022 09:25:00 - INFO - codeparrot_training - Step 14411: {'lr': 0.000421958041321445, 'samples': 2767104, 'steps': 14411, 'loss/train': 0.4013863205909729} +01/27/2022 09:25:03 - INFO - codeparrot_training - Step 14412: {'lr': 0.0004219461639369281, 'samples': 2767296, 'steps': 14412, 'loss/train': 0.5333157330751419} +01/27/2022 09:25:06 - INFO - codeparrot_training - Step 14413: {'lr': 0.0004219342858158485, 'samples': 2767488, 'steps': 14413, 'loss/train': 0.8734095096588135} +01/27/2022 09:25:10 - INFO - codeparrot_training - Step 14414: {'lr': 0.000421922406958257, 'samples': 2767680, 'steps': 14414, 'loss/train': 0.595910981297493} +01/27/2022 09:25:13 - INFO - codeparrot_training - Step 14415: {'lr': 0.00042191052736420445, 'samples': 2767872, 'steps': 14415, 'loss/train': 0.9594951868057251} +01/27/2022 09:25:16 - INFO - codeparrot_training - Step 14416: {'lr': 0.0004218986470337419, 'samples': 2768064, 'steps': 14416, 'loss/train': 1.3609235286712646} +01/27/2022 09:25:19 - INFO - codeparrot_training - Step 14417: {'lr': 0.00042188676596692, 'samples': 2768256, 'steps': 14417, 'loss/train': 1.1135479509830475} +01/27/2022 09:25:22 - INFO - codeparrot_training - Step 14418: {'lr': 0.0004218748841637899, 'samples': 2768448, 'steps': 14418, 'loss/train': 0.7475330829620361} +01/27/2022 09:25:27 - INFO - codeparrot_training - Step 14419: {'lr': 0.0004218630016244023, 'samples': 2768640, 'steps': 14419, 'loss/train': 1.1087610125541687} +01/27/2022 09:25:30 - INFO - codeparrot_training - Step 14420: {'lr': 0.0004218511183488082, 'samples': 2768832, 'steps': 14420, 'loss/train': 0.6970864981412888} +01/27/2022 09:25:33 - INFO - codeparrot_training - Step 14421: {'lr': 0.0004218392343370584, 'samples': 2769024, 'steps': 14421, 'loss/train': 1.0141584277153015} +01/27/2022 09:25:36 - INFO - codeparrot_training - Step 14422: {'lr': 0.000421827349589204, 'samples': 2769216, 'steps': 14422, 'loss/train': 0.8268519937992096} +01/27/2022 09:25:39 - INFO - codeparrot_training - Step 14423: {'lr': 0.0004218154641052957, 'samples': 2769408, 'steps': 14423, 'loss/train': 0.7507629096508026} +01/27/2022 09:25:42 - INFO - codeparrot_training - Step 14424: {'lr': 0.0004218035778853846, 'samples': 2769600, 'steps': 14424, 'loss/train': 0.5817403346300125} +01/27/2022 09:25:46 - INFO - codeparrot_training - Step 14425: {'lr': 0.0004217916909295215, 'samples': 2769792, 'steps': 14425, 'loss/train': 0.8972252011299133} +01/27/2022 09:25:49 - INFO - codeparrot_training - Step 14426: {'lr': 0.00042177980323775734, 'samples': 2769984, 'steps': 14426, 'loss/train': 0.8555029034614563} +01/27/2022 09:25:52 - INFO - codeparrot_training - Step 14427: {'lr': 0.00042176791481014303, 'samples': 2770176, 'steps': 14427, 'loss/train': 0.913575142621994} +01/27/2022 09:25:56 - INFO - codeparrot_training - Step 14428: {'lr': 0.0004217560256467295, 'samples': 2770368, 'steps': 14428, 'loss/train': 0.6797819137573242} +01/27/2022 09:25:59 - INFO - codeparrot_training - Step 14429: {'lr': 0.00042174413574756775, 'samples': 2770560, 'steps': 14429, 'loss/train': 1.1368997991085052} +01/27/2022 09:26:03 - INFO - codeparrot_training - Step 14430: {'lr': 0.0004217322451127086, 'samples': 2770752, 'steps': 14430, 'loss/train': 0.9026723206043243} +01/27/2022 09:26:06 - INFO - codeparrot_training - Step 14431: {'lr': 0.00042172035374220306, 'samples': 2770944, 'steps': 14431, 'loss/train': 1.1765846908092499} +01/27/2022 09:26:09 - INFO - codeparrot_training - Step 14432: {'lr': 0.0004217084616361021, 'samples': 2771136, 'steps': 14432, 'loss/train': 0.6705069243907928} +01/27/2022 09:26:12 - INFO - codeparrot_training - Step 14433: {'lr': 0.00042169656879445657, 'samples': 2771328, 'steps': 14433, 'loss/train': 0.6782628893852234} +01/27/2022 09:26:15 - INFO - codeparrot_training - Step 14434: {'lr': 0.00042168467521731747, 'samples': 2771520, 'steps': 14434, 'loss/train': 0.42917583882808685} +01/27/2022 09:26:18 - INFO - codeparrot_training - Step 14435: {'lr': 0.00042167278090473573, 'samples': 2771712, 'steps': 14435, 'loss/train': 0.47453588247299194} +01/27/2022 09:26:24 - INFO - codeparrot_training - Step 14436: {'lr': 0.0004216608858567623, 'samples': 2771904, 'steps': 14436, 'loss/train': 1.1609399020671844} +01/27/2022 09:26:28 - INFO - codeparrot_training - Step 14437: {'lr': 0.00042164899007344814, 'samples': 2772096, 'steps': 14437, 'loss/train': 4.625206232070923} +01/27/2022 09:26:31 - INFO - codeparrot_training - Step 14438: {'lr': 0.00042163709355484425, 'samples': 2772288, 'steps': 14438, 'loss/train': 0.03907719813287258} +01/27/2022 09:26:34 - INFO - codeparrot_training - Step 14439: {'lr': 0.0004216251963010015, 'samples': 2772480, 'steps': 14439, 'loss/train': 0.898187130689621} +01/27/2022 09:26:37 - INFO - codeparrot_training - Step 14440: {'lr': 0.0004216132983119709, 'samples': 2772672, 'steps': 14440, 'loss/train': 1.130095660686493} +01/27/2022 09:26:40 - INFO - codeparrot_training - Step 14441: {'lr': 0.00042160139958780346, 'samples': 2772864, 'steps': 14441, 'loss/train': 0.3290546089410782} +01/27/2022 09:26:43 - INFO - codeparrot_training - Step 14442: {'lr': 0.0004215895001285501, 'samples': 2773056, 'steps': 14442, 'loss/train': 0.5484882295131683} +01/27/2022 09:26:47 - INFO - codeparrot_training - Step 14443: {'lr': 0.0004215775999342618, 'samples': 2773248, 'steps': 14443, 'loss/train': 1.0173275470733643} +01/27/2022 09:26:50 - INFO - codeparrot_training - Step 14444: {'lr': 0.0004215656990049896, 'samples': 2773440, 'steps': 14444, 'loss/train': 0.93079474568367} +01/27/2022 09:26:54 - INFO - codeparrot_training - Step 14445: {'lr': 0.0004215537973407844, 'samples': 2773632, 'steps': 14445, 'loss/train': 0.810340404510498} +01/27/2022 09:26:57 - INFO - codeparrot_training - Step 14446: {'lr': 0.0004215418949416972, 'samples': 2773824, 'steps': 14446, 'loss/train': 0.8437301516532898} +01/27/2022 09:27:00 - INFO - codeparrot_training - Step 14447: {'lr': 0.00042152999180777894, 'samples': 2774016, 'steps': 14447, 'loss/train': 0.6909996271133423} +01/27/2022 09:27:04 - INFO - codeparrot_training - Step 14448: {'lr': 0.0004215180879390807, 'samples': 2774208, 'steps': 14448, 'loss/train': 0.7857309579849243} +01/27/2022 09:27:07 - INFO - codeparrot_training - Step 14449: {'lr': 0.0004215061833356535, 'samples': 2774400, 'steps': 14449, 'loss/train': 0.6656115800142288} +01/27/2022 09:27:10 - INFO - codeparrot_training - Step 14450: {'lr': 0.00042149427799754817, 'samples': 2774592, 'steps': 14450, 'loss/train': 0.7264866828918457} +01/27/2022 09:27:13 - INFO - codeparrot_training - Step 14451: {'lr': 0.00042148237192481586, 'samples': 2774784, 'steps': 14451, 'loss/train': 1.5580299496650696} +01/27/2022 09:27:16 - INFO - codeparrot_training - Step 14452: {'lr': 0.0004214704651175075, 'samples': 2774976, 'steps': 14452, 'loss/train': 0.9654604196548462} +01/27/2022 09:27:19 - INFO - codeparrot_training - Step 14453: {'lr': 0.0004214585575756742, 'samples': 2775168, 'steps': 14453, 'loss/train': 2.9237101078033447} +01/27/2022 09:27:24 - INFO - codeparrot_training - Step 14454: {'lr': 0.0004214466492993668, 'samples': 2775360, 'steps': 14454, 'loss/train': 1.1588982939720154} +01/27/2022 09:27:27 - INFO - codeparrot_training - Step 14455: {'lr': 0.00042143474028863637, 'samples': 2775552, 'steps': 14455, 'loss/train': 0.8973933756351471} +01/27/2022 09:27:30 - INFO - codeparrot_training - Step 14456: {'lr': 0.000421422830543534, 'samples': 2775744, 'steps': 14456, 'loss/train': 1.8617772459983826} +01/27/2022 09:27:34 - INFO - codeparrot_training - Step 14457: {'lr': 0.0004214109200641106, 'samples': 2775936, 'steps': 14457, 'loss/train': 0.6924179792404175} +01/27/2022 09:27:37 - INFO - codeparrot_training - Step 14458: {'lr': 0.00042139900885041734, 'samples': 2776128, 'steps': 14458, 'loss/train': 1.3103719353675842} +01/27/2022 09:27:40 - INFO - codeparrot_training - Step 14459: {'lr': 0.00042138709690250507, 'samples': 2776320, 'steps': 14459, 'loss/train': 0.9275031387805939} +01/27/2022 09:27:43 - INFO - codeparrot_training - Step 14460: {'lr': 0.0004213751842204249, 'samples': 2776512, 'steps': 14460, 'loss/train': 1.241137146949768} +01/27/2022 09:27:46 - INFO - codeparrot_training - Step 14461: {'lr': 0.00042136327080422785, 'samples': 2776704, 'steps': 14461, 'loss/train': 0.41795937716960907} +01/27/2022 09:27:49 - INFO - codeparrot_training - Step 14462: {'lr': 0.0004213513566539651, 'samples': 2776896, 'steps': 14462, 'loss/train': 0.7933521866798401} +01/27/2022 09:27:55 - INFO - codeparrot_training - Step 14463: {'lr': 0.0004213394417696874, 'samples': 2777088, 'steps': 14463, 'loss/train': 0.9667510092258453} +01/27/2022 09:27:59 - INFO - codeparrot_training - Step 14464: {'lr': 0.00042132752615144597, 'samples': 2777280, 'steps': 14464, 'loss/train': 0.8516510725021362} +01/27/2022 09:28:02 - INFO - codeparrot_training - Step 14465: {'lr': 0.00042131560979929186, 'samples': 2777472, 'steps': 14465, 'loss/train': 1.408642441034317} +01/27/2022 09:28:05 - INFO - codeparrot_training - Step 14466: {'lr': 0.00042130369271327605, 'samples': 2777664, 'steps': 14466, 'loss/train': 0.9582488536834717} +01/27/2022 09:28:08 - INFO - codeparrot_training - Step 14467: {'lr': 0.0004212917748934496, 'samples': 2777856, 'steps': 14467, 'loss/train': 0.5621844381093979} +01/27/2022 09:28:11 - INFO - codeparrot_training - Step 14468: {'lr': 0.00042127985633986365, 'samples': 2778048, 'steps': 14468, 'loss/train': 0.8498403131961823} +01/27/2022 09:28:14 - INFO - codeparrot_training - Step 14469: {'lr': 0.00042126793705256913, 'samples': 2778240, 'steps': 14469, 'loss/train': 1.3029142320156097} +01/27/2022 09:28:17 - INFO - codeparrot_training - Step 14470: {'lr': 0.00042125601703161706, 'samples': 2778432, 'steps': 14470, 'loss/train': 0.8963984549045563} +01/27/2022 09:28:21 - INFO - codeparrot_training - Step 14471: {'lr': 0.00042124409627705873, 'samples': 2778624, 'steps': 14471, 'loss/train': 1.1731768548488617} +01/27/2022 09:28:25 - INFO - codeparrot_training - Step 14472: {'lr': 0.00042123217478894504, 'samples': 2778816, 'steps': 14472, 'loss/train': 0.9095066785812378} +01/27/2022 09:28:28 - INFO - codeparrot_training - Step 14473: {'lr': 0.0004212202525673271, 'samples': 2779008, 'steps': 14473, 'loss/train': 0.5193129926919937} +01/27/2022 09:28:31 - INFO - codeparrot_training - Step 14474: {'lr': 0.00042120832961225585, 'samples': 2779200, 'steps': 14474, 'loss/train': 0.9703179001808167} +01/27/2022 09:28:35 - INFO - codeparrot_training - Step 14475: {'lr': 0.00042119640592378263, 'samples': 2779392, 'steps': 14475, 'loss/train': 0.0939881019294262} +01/27/2022 09:28:38 - INFO - codeparrot_training - Step 14476: {'lr': 0.00042118448150195827, 'samples': 2779584, 'steps': 14476, 'loss/train': 0.627482995390892} +01/27/2022 09:28:41 - INFO - codeparrot_training - Step 14477: {'lr': 0.000421172556346834, 'samples': 2779776, 'steps': 14477, 'loss/train': 0.5072409063577652} +01/27/2022 09:28:44 - INFO - codeparrot_training - Step 14478: {'lr': 0.00042116063045846073, 'samples': 2779968, 'steps': 14478, 'loss/train': 0.9221839606761932} +01/27/2022 09:28:47 - INFO - codeparrot_training - Step 14479: {'lr': 0.00042114870383688985, 'samples': 2780160, 'steps': 14479, 'loss/train': 0.7872251272201538} +01/27/2022 09:28:50 - INFO - codeparrot_training - Step 14480: {'lr': 0.0004211367764821722, 'samples': 2780352, 'steps': 14480, 'loss/train': 1.4597254693508148} +01/27/2022 09:28:55 - INFO - codeparrot_training - Step 14481: {'lr': 0.00042112484839435893, 'samples': 2780544, 'steps': 14481, 'loss/train': 0.7884701192378998} +01/27/2022 09:28:58 - INFO - codeparrot_training - Step 14482: {'lr': 0.00042111291957350113, 'samples': 2780736, 'steps': 14482, 'loss/train': 0.9117384552955627} +01/27/2022 09:29:01 - INFO - codeparrot_training - Step 14483: {'lr': 0.00042110099001964996, 'samples': 2780928, 'steps': 14483, 'loss/train': 0.7409770041704178} +01/27/2022 09:29:04 - INFO - codeparrot_training - Step 14484: {'lr': 0.0004210890597328564, 'samples': 2781120, 'steps': 14484, 'loss/train': 1.0366149544715881} +01/27/2022 09:29:07 - INFO - codeparrot_training - Step 14485: {'lr': 0.0004210771287131717, 'samples': 2781312, 'steps': 14485, 'loss/train': 1.0737177729606628} +01/27/2022 09:29:10 - INFO - codeparrot_training - Step 14486: {'lr': 0.00042106519696064694, 'samples': 2781504, 'steps': 14486, 'loss/train': 0.9778135120868683} +01/27/2022 09:29:14 - INFO - codeparrot_training - Step 14487: {'lr': 0.0004210532644753331, 'samples': 2781696, 'steps': 14487, 'loss/train': 0.5547505617141724} +01/27/2022 09:29:17 - INFO - codeparrot_training - Step 14488: {'lr': 0.00042104133125728146, 'samples': 2781888, 'steps': 14488, 'loss/train': 0.9681587219238281} +01/27/2022 09:29:23 - INFO - codeparrot_training - Step 14489: {'lr': 0.00042102939730654304, 'samples': 2782080, 'steps': 14489, 'loss/train': 0.978799045085907} +01/27/2022 09:29:26 - INFO - codeparrot_training - Step 14490: {'lr': 0.000421017462623169, 'samples': 2782272, 'steps': 14490, 'loss/train': 0.4087405353784561} +01/27/2022 09:29:29 - INFO - codeparrot_training - Step 14491: {'lr': 0.0004210055272072104, 'samples': 2782464, 'steps': 14491, 'loss/train': 0.8553954362869263} +01/27/2022 09:29:32 - INFO - codeparrot_training - Step 14492: {'lr': 0.00042099359105871856, 'samples': 2782656, 'steps': 14492, 'loss/train': 0.7905256748199463} +01/27/2022 09:29:36 - INFO - codeparrot_training - Step 14493: {'lr': 0.0004209816541777444, 'samples': 2782848, 'steps': 14493, 'loss/train': 1.2241169214248657} +01/27/2022 09:29:39 - INFO - codeparrot_training - Step 14494: {'lr': 0.0004209697165643391, 'samples': 2783040, 'steps': 14494, 'loss/train': 1.6836492419242859} +01/27/2022 09:29:42 - INFO - codeparrot_training - Step 14495: {'lr': 0.0004209577782185538, 'samples': 2783232, 'steps': 14495, 'loss/train': 1.0210855901241302} +01/27/2022 09:29:45 - INFO - codeparrot_training - Step 14496: {'lr': 0.0004209458391404397, 'samples': 2783424, 'steps': 14496, 'loss/train': 0.6866628974676132} +01/27/2022 09:29:48 - INFO - codeparrot_training - Step 14497: {'lr': 0.0004209338993300479, 'samples': 2783616, 'steps': 14497, 'loss/train': 0.98428013920784} +01/27/2022 09:29:51 - INFO - codeparrot_training - Step 14498: {'lr': 0.00042092195878742954, 'samples': 2783808, 'steps': 14498, 'loss/train': 0.6694585382938385} +01/27/2022 09:29:56 - INFO - codeparrot_training - Step 14499: {'lr': 0.0004209100175126358, 'samples': 2784000, 'steps': 14499, 'loss/train': 0.6646627485752106} +01/27/2022 09:29:59 - INFO - codeparrot_training - Step 14500: {'lr': 0.0004208980755057178, 'samples': 2784192, 'steps': 14500, 'loss/train': 0.7417833209037781} +01/27/2022 09:30:02 - INFO - codeparrot_training - Step 14501: {'lr': 0.0004208861327667268, 'samples': 2784384, 'steps': 14501, 'loss/train': 1.2180006802082062} +01/27/2022 09:30:05 - INFO - codeparrot_training - Step 14502: {'lr': 0.00042087418929571377, 'samples': 2784576, 'steps': 14502, 'loss/train': 1.1603479385375977} +01/27/2022 09:30:08 - INFO - codeparrot_training - Step 14503: {'lr': 0.00042086224509272995, 'samples': 2784768, 'steps': 14503, 'loss/train': 1.1282957196235657} +01/27/2022 09:30:11 - INFO - codeparrot_training - Step 14504: {'lr': 0.0004208503001578266, 'samples': 2784960, 'steps': 14504, 'loss/train': 0.7658489942550659} +01/27/2022 09:30:15 - INFO - codeparrot_training - Step 14505: {'lr': 0.00042083835449105477, 'samples': 2785152, 'steps': 14505, 'loss/train': 0.9246900379657745} +01/27/2022 09:30:18 - INFO - codeparrot_training - Step 14506: {'lr': 0.00042082640809246576, 'samples': 2785344, 'steps': 14506, 'loss/train': 1.0561244487762451} +01/27/2022 09:30:22 - INFO - codeparrot_training - Step 14507: {'lr': 0.0004208144609621106, 'samples': 2785536, 'steps': 14507, 'loss/train': 0.3948938548564911} +01/27/2022 09:30:25 - INFO - codeparrot_training - Step 14508: {'lr': 0.0004208025131000405, 'samples': 2785728, 'steps': 14508, 'loss/train': 0.9509541392326355} +01/27/2022 09:30:29 - INFO - codeparrot_training - Step 14509: {'lr': 0.0004207905645063067, 'samples': 2785920, 'steps': 14509, 'loss/train': 0.7590352892875671} +01/27/2022 09:30:32 - INFO - codeparrot_training - Step 14510: {'lr': 0.00042077861518096033, 'samples': 2786112, 'steps': 14510, 'loss/train': 0.2022121176123619} +01/27/2022 09:30:35 - INFO - codeparrot_training - Step 14511: {'lr': 0.0004207666651240526, 'samples': 2786304, 'steps': 14511, 'loss/train': 0.4118529260158539} +01/27/2022 09:30:38 - INFO - codeparrot_training - Step 14512: {'lr': 0.0004207547143356347, 'samples': 2786496, 'steps': 14512, 'loss/train': 0.6770116835832596} +01/27/2022 09:30:41 - INFO - codeparrot_training - Step 14513: {'lr': 0.00042074276281575787, 'samples': 2786688, 'steps': 14513, 'loss/train': 1.024745911359787} +01/27/2022 09:30:44 - INFO - codeparrot_training - Step 14514: {'lr': 0.00042073081056447325, 'samples': 2786880, 'steps': 14514, 'loss/train': 0.9472421407699585} +01/27/2022 09:30:48 - INFO - codeparrot_training - Step 14515: {'lr': 0.00042071885758183204, 'samples': 2787072, 'steps': 14515, 'loss/train': 0.4195275753736496} +01/27/2022 09:30:54 - INFO - codeparrot_training - Step 14516: {'lr': 0.00042070690386788545, 'samples': 2787264, 'steps': 14516, 'loss/train': 0.6508060991764069} +01/27/2022 09:30:57 - INFO - codeparrot_training - Step 14517: {'lr': 0.0004206949494226847, 'samples': 2787456, 'steps': 14517, 'loss/train': 0.6054255813360214} +01/27/2022 09:31:00 - INFO - codeparrot_training - Step 14518: {'lr': 0.000420682994246281, 'samples': 2787648, 'steps': 14518, 'loss/train': 1.2019346058368683} +01/27/2022 09:31:04 - INFO - codeparrot_training - Step 14519: {'lr': 0.00042067103833872554, 'samples': 2787840, 'steps': 14519, 'loss/train': 0.7660499811172485} +01/27/2022 09:31:07 - INFO - codeparrot_training - Step 14520: {'lr': 0.0004206590817000695, 'samples': 2788032, 'steps': 14520, 'loss/train': 0.906202107667923} +01/27/2022 09:31:10 - INFO - codeparrot_training - Step 14521: {'lr': 0.0004206471243303642, 'samples': 2788224, 'steps': 14521, 'loss/train': 1.327378660440445} +01/27/2022 09:31:13 - INFO - codeparrot_training - Step 14522: {'lr': 0.0004206351662296608, 'samples': 2788416, 'steps': 14522, 'loss/train': 1.3098507821559906} +01/27/2022 09:31:16 - INFO - codeparrot_training - Step 14523: {'lr': 0.0004206232073980105, 'samples': 2788608, 'steps': 14523, 'loss/train': 0.9097177684307098} +01/27/2022 09:31:19 - INFO - codeparrot_training - Step 14524: {'lr': 0.00042061124783546454, 'samples': 2788800, 'steps': 14524, 'loss/train': 0.5270503163337708} +01/27/2022 09:31:23 - INFO - codeparrot_training - Step 14525: {'lr': 0.0004205992875420742, 'samples': 2788992, 'steps': 14525, 'loss/train': 1.0339031517505646} +01/27/2022 09:31:27 - INFO - codeparrot_training - Step 14526: {'lr': 0.0004205873265178907, 'samples': 2789184, 'steps': 14526, 'loss/train': 0.939132571220398} +01/27/2022 09:31:30 - INFO - codeparrot_training - Step 14527: {'lr': 0.0004205753647629653, 'samples': 2789376, 'steps': 14527, 'loss/train': 0.28479308634996414} +01/27/2022 09:31:33 - INFO - codeparrot_training - Step 14528: {'lr': 0.0004205634022773491, 'samples': 2789568, 'steps': 14528, 'loss/train': 0.9555148780345917} +01/27/2022 09:31:36 - INFO - codeparrot_training - Step 14529: {'lr': 0.0004205514390610935, 'samples': 2789760, 'steps': 14529, 'loss/train': 0.4125848561525345} +01/27/2022 09:31:40 - INFO - codeparrot_training - Step 14530: {'lr': 0.00042053947511424975, 'samples': 2789952, 'steps': 14530, 'loss/train': 0.9416978359222412} +01/27/2022 09:31:43 - INFO - codeparrot_training - Step 14531: {'lr': 0.00042052751043686895, 'samples': 2790144, 'steps': 14531, 'loss/train': 1.3614908158779144} +01/27/2022 09:31:46 - INFO - codeparrot_training - Step 14532: {'lr': 0.00042051554502900245, 'samples': 2790336, 'steps': 14532, 'loss/train': 1.0034666955471039} +01/27/2022 09:31:49 - INFO - codeparrot_training - Step 14533: {'lr': 0.0004205035788907015, 'samples': 2790528, 'steps': 14533, 'loss/train': 0.5220656991004944} +01/27/2022 09:31:52 - INFO - codeparrot_training - Step 14534: {'lr': 0.0004204916120220174, 'samples': 2790720, 'steps': 14534, 'loss/train': 0.19978545606136322} +01/27/2022 09:31:56 - INFO - codeparrot_training - Step 14535: {'lr': 0.00042047964442300137, 'samples': 2790912, 'steps': 14535, 'loss/train': 1.1870808899402618} +01/27/2022 09:32:00 - INFO - codeparrot_training - Step 14536: {'lr': 0.0004204676760937046, 'samples': 2791104, 'steps': 14536, 'loss/train': 0.5705276727676392} +01/27/2022 09:32:03 - INFO - codeparrot_training - Step 14537: {'lr': 0.00042045570703417857, 'samples': 2791296, 'steps': 14537, 'loss/train': 0.8454370200634003} +01/27/2022 09:32:06 - INFO - codeparrot_training - Step 14538: {'lr': 0.00042044373724447434, 'samples': 2791488, 'steps': 14538, 'loss/train': 1.0435466766357422} +01/27/2022 09:32:09 - INFO - codeparrot_training - Step 14539: {'lr': 0.0004204317667246432, 'samples': 2791680, 'steps': 14539, 'loss/train': 0.621541902422905} +01/27/2022 09:32:12 - INFO - codeparrot_training - Step 14540: {'lr': 0.00042041979547473665, 'samples': 2791872, 'steps': 14540, 'loss/train': 0.9140615165233612} +01/27/2022 09:32:15 - INFO - codeparrot_training - Step 14541: {'lr': 0.0004204078234948057, 'samples': 2792064, 'steps': 14541, 'loss/train': 1.0425889492034912} +01/27/2022 09:32:18 - INFO - codeparrot_training - Step 14542: {'lr': 0.00042039585078490173, 'samples': 2792256, 'steps': 14542, 'loss/train': 0.8605214059352875} +01/27/2022 09:32:22 - INFO - codeparrot_training - Step 14543: {'lr': 0.000420383877345076, 'samples': 2792448, 'steps': 14543, 'loss/train': 0.9905081391334534} +01/27/2022 09:32:28 - INFO - codeparrot_training - Step 14544: {'lr': 0.00042037190317538, 'samples': 2792640, 'steps': 14544, 'loss/train': 0.6540424525737762} +01/27/2022 09:32:31 - INFO - codeparrot_training - Step 14545: {'lr': 0.00042035992827586474, 'samples': 2792832, 'steps': 14545, 'loss/train': 1.0332371592521667} +01/27/2022 09:32:34 - INFO - codeparrot_training - Step 14546: {'lr': 0.00042034795264658163, 'samples': 2793024, 'steps': 14546, 'loss/train': 1.0012752413749695} +01/27/2022 09:32:37 - INFO - codeparrot_training - Step 14547: {'lr': 0.00042033597628758206, 'samples': 2793216, 'steps': 14547, 'loss/train': 0.7677215337753296} +01/27/2022 09:32:40 - INFO - codeparrot_training - Step 14548: {'lr': 0.00042032399919891724, 'samples': 2793408, 'steps': 14548, 'loss/train': 0.5619672685861588} +01/27/2022 09:32:44 - INFO - codeparrot_training - Step 14549: {'lr': 0.0004203120213806385, 'samples': 2793600, 'steps': 14549, 'loss/train': 0.9811786115169525} +01/27/2022 09:32:47 - INFO - codeparrot_training - Step 14550: {'lr': 0.0004203000428327971, 'samples': 2793792, 'steps': 14550, 'loss/train': 0.8689026832580566} +01/27/2022 09:32:50 - INFO - codeparrot_training - Step 14551: {'lr': 0.00042028806355544443, 'samples': 2793984, 'steps': 14551, 'loss/train': 1.085404872894287} +01/27/2022 09:32:54 - INFO - codeparrot_training - Step 14552: {'lr': 0.0004202760835486317, 'samples': 2794176, 'steps': 14552, 'loss/train': 0.9343622624874115} +01/27/2022 09:32:58 - INFO - codeparrot_training - Step 14553: {'lr': 0.00042026410281241033, 'samples': 2794368, 'steps': 14553, 'loss/train': 0.2766171544790268} +01/27/2022 09:33:01 - INFO - codeparrot_training - Step 14554: {'lr': 0.00042025212134683165, 'samples': 2794560, 'steps': 14554, 'loss/train': 1.3742137849330902} +01/27/2022 09:33:04 - INFO - codeparrot_training - Step 14555: {'lr': 0.0004202401391519469, 'samples': 2794752, 'steps': 14555, 'loss/train': 1.0953112542629242} +01/27/2022 09:33:07 - INFO - codeparrot_training - Step 14556: {'lr': 0.0004202281562278075, 'samples': 2794944, 'steps': 14556, 'loss/train': 0.50103360414505} +01/27/2022 09:33:10 - INFO - codeparrot_training - Step 14557: {'lr': 0.0004202161725744647, 'samples': 2795136, 'steps': 14557, 'loss/train': 1.1619705855846405} +01/27/2022 09:33:13 - INFO - codeparrot_training - Step 14558: {'lr': 0.0004202041881919699, 'samples': 2795328, 'steps': 14558, 'loss/train': 0.49301183223724365} +01/27/2022 09:33:16 - INFO - codeparrot_training - Step 14559: {'lr': 0.0004201922030803743, 'samples': 2795520, 'steps': 14559, 'loss/train': 0.48405472934246063} +01/27/2022 09:33:20 - INFO - codeparrot_training - Step 14560: {'lr': 0.0004201802172397295, 'samples': 2795712, 'steps': 14560, 'loss/train': 0.6241238862276077} +01/27/2022 09:33:26 - INFO - codeparrot_training - Step 14561: {'lr': 0.0004201682306700866, 'samples': 2795904, 'steps': 14561, 'loss/train': 0.9935800731182098} +01/27/2022 09:33:29 - INFO - codeparrot_training - Step 14562: {'lr': 0.00042015624337149703, 'samples': 2796096, 'steps': 14562, 'loss/train': 1.5448240041732788} +01/27/2022 09:33:32 - INFO - codeparrot_training - Step 14563: {'lr': 0.0004201442553440121, 'samples': 2796288, 'steps': 14563, 'loss/train': 0.949792742729187} +01/27/2022 09:33:35 - INFO - codeparrot_training - Step 14564: {'lr': 0.00042013226658768333, 'samples': 2796480, 'steps': 14564, 'loss/train': 0.8796691596508026} +01/27/2022 09:33:38 - INFO - codeparrot_training - Step 14565: {'lr': 0.0004201202771025618, 'samples': 2796672, 'steps': 14565, 'loss/train': 1.0586282908916473} +01/27/2022 09:33:41 - INFO - codeparrot_training - Step 14566: {'lr': 0.0004201082868886992, 'samples': 2796864, 'steps': 14566, 'loss/train': 0.703694298863411} +01/27/2022 09:33:44 - INFO - codeparrot_training - Step 14567: {'lr': 0.00042009629594614656, 'samples': 2797056, 'steps': 14567, 'loss/train': 0.49180278182029724} +01/27/2022 09:33:48 - INFO - codeparrot_training - Step 14568: {'lr': 0.0004200843042749555, 'samples': 2797248, 'steps': 14568, 'loss/train': 0.923474907875061} +01/27/2022 09:33:51 - INFO - codeparrot_training - Step 14569: {'lr': 0.0004200723118751772, 'samples': 2797440, 'steps': 14569, 'loss/train': 0.8950442969799042} +01/27/2022 09:33:55 - INFO - codeparrot_training - Step 14570: {'lr': 0.00042006031874686315, 'samples': 2797632, 'steps': 14570, 'loss/train': 0.7863157689571381} +01/27/2022 09:33:58 - INFO - codeparrot_training - Step 14571: {'lr': 0.00042004832489006474, 'samples': 2797824, 'steps': 14571, 'loss/train': 0.07356058806180954} +01/27/2022 09:34:01 - INFO - codeparrot_training - Step 14572: {'lr': 0.0004200363303048332, 'samples': 2798016, 'steps': 14572, 'loss/train': 0.9502094686031342} +01/27/2022 09:34:04 - INFO - codeparrot_training - Step 14573: {'lr': 0.00042002433499122016, 'samples': 2798208, 'steps': 14573, 'loss/train': 0.5144761651754379} +01/27/2022 09:34:08 - INFO - codeparrot_training - Step 14574: {'lr': 0.00042001233894927684, 'samples': 2798400, 'steps': 14574, 'loss/train': 0.557689368724823} +01/27/2022 09:34:11 - INFO - codeparrot_training - Step 14575: {'lr': 0.0004200003421790546, 'samples': 2798592, 'steps': 14575, 'loss/train': 0.7439205944538116} +01/27/2022 09:34:14 - INFO - codeparrot_training - Step 14576: {'lr': 0.0004199883446806048, 'samples': 2798784, 'steps': 14576, 'loss/train': 1.2817760109901428} +01/27/2022 09:34:17 - INFO - codeparrot_training - Step 14577: {'lr': 0.00041997634645397897, 'samples': 2798976, 'steps': 14577, 'loss/train': 1.0155308246612549} +01/27/2022 09:34:20 - INFO - codeparrot_training - Step 14578: {'lr': 0.0004199643474992285, 'samples': 2799168, 'steps': 14578, 'loss/train': 0.6829057037830353} +01/27/2022 09:34:25 - INFO - codeparrot_training - Step 14579: {'lr': 0.00041995234781640466, 'samples': 2799360, 'steps': 14579, 'loss/train': 0.533542737364769} +01/27/2022 09:34:28 - INFO - codeparrot_training - Step 14580: {'lr': 0.00041994034740555896, 'samples': 2799552, 'steps': 14580, 'loss/train': 0.8295055031776428} +01/27/2022 09:34:32 - INFO - codeparrot_training - Step 14581: {'lr': 0.00041992834626674273, 'samples': 2799744, 'steps': 14581, 'loss/train': 1.362406611442566} +01/27/2022 09:34:35 - INFO - codeparrot_training - Step 14582: {'lr': 0.0004199163444000075, 'samples': 2799936, 'steps': 14582, 'loss/train': 0.8247813284397125} +01/27/2022 09:34:38 - INFO - codeparrot_training - Step 14583: {'lr': 0.00041990434180540453, 'samples': 2800128, 'steps': 14583, 'loss/train': 1.0878243148326874} +01/27/2022 09:34:41 - INFO - codeparrot_training - Step 14584: {'lr': 0.00041989233848298534, 'samples': 2800320, 'steps': 14584, 'loss/train': 0.8054029047489166} +01/27/2022 09:34:44 - INFO - codeparrot_training - Step 14585: {'lr': 0.00041988033443280136, 'samples': 2800512, 'steps': 14585, 'loss/train': 0.7279120534658432} +01/27/2022 09:34:47 - INFO - codeparrot_training - Step 14586: {'lr': 0.00041986832965490396, 'samples': 2800704, 'steps': 14586, 'loss/train': 0.6369442641735077} +01/27/2022 09:34:53 - INFO - codeparrot_training - Step 14587: {'lr': 0.0004198563241493445, 'samples': 2800896, 'steps': 14587, 'loss/train': 0.9598861634731293} +01/27/2022 09:34:57 - INFO - codeparrot_training - Step 14588: {'lr': 0.00041984431791617456, 'samples': 2801088, 'steps': 14588, 'loss/train': 0.9072262644767761} +01/27/2022 09:35:00 - INFO - codeparrot_training - Step 14589: {'lr': 0.00041983231095544545, 'samples': 2801280, 'steps': 14589, 'loss/train': 0.8308858573436737} +01/27/2022 09:35:03 - INFO - codeparrot_training - Step 14590: {'lr': 0.00041982030326720866, 'samples': 2801472, 'steps': 14590, 'loss/train': 0.773174375295639} +01/27/2022 09:35:06 - INFO - codeparrot_training - Step 14591: {'lr': 0.00041980829485151563, 'samples': 2801664, 'steps': 14591, 'loss/train': 0.7474144399166107} +01/27/2022 09:35:09 - INFO - codeparrot_training - Step 14592: {'lr': 0.00041979628570841776, 'samples': 2801856, 'steps': 14592, 'loss/train': 1.0304082334041595} +01/27/2022 09:35:12 - INFO - codeparrot_training - Step 14593: {'lr': 0.00041978427583796654, 'samples': 2802048, 'steps': 14593, 'loss/train': 0.8465198278427124} +01/27/2022 09:35:15 - INFO - codeparrot_training - Step 14594: {'lr': 0.00041977226524021337, 'samples': 2802240, 'steps': 14594, 'loss/train': 0.9279935359954834} +01/27/2022 09:35:19 - INFO - codeparrot_training - Step 14595: {'lr': 0.0004197602539152098, 'samples': 2802432, 'steps': 14595, 'loss/train': 1.1681474447250366} +01/27/2022 09:35:23 - INFO - codeparrot_training - Step 14596: {'lr': 0.00041974824186300706, 'samples': 2802624, 'steps': 14596, 'loss/train': 0.606344148516655} +01/27/2022 09:35:26 - INFO - codeparrot_training - Step 14597: {'lr': 0.0004197362290836569, 'samples': 2802816, 'steps': 14597, 'loss/train': 0.962935209274292} +01/27/2022 09:35:29 - INFO - codeparrot_training - Step 14598: {'lr': 0.00041972421557721055, 'samples': 2803008, 'steps': 14598, 'loss/train': 0.7736509144306183} +01/27/2022 09:35:32 - INFO - codeparrot_training - Step 14599: {'lr': 0.00041971220134371957, 'samples': 2803200, 'steps': 14599, 'loss/train': 0.9043280482292175} +01/27/2022 09:35:36 - INFO - codeparrot_training - Step 14600: {'lr': 0.00041970018638323546, 'samples': 2803392, 'steps': 14600, 'loss/train': 0.7137068063020706} +01/27/2022 09:35:39 - INFO - codeparrot_training - Step 14601: {'lr': 0.0004196881706958096, 'samples': 2803584, 'steps': 14601, 'loss/train': 0.5279713422060013} +01/27/2022 09:35:42 - INFO - codeparrot_training - Step 14602: {'lr': 0.00041967615428149346, 'samples': 2803776, 'steps': 14602, 'loss/train': 0.9779272377490997} +01/27/2022 09:35:45 - INFO - codeparrot_training - Step 14603: {'lr': 0.0004196641371403386, 'samples': 2803968, 'steps': 14603, 'loss/train': 0.9970919787883759} +01/27/2022 09:35:48 - INFO - codeparrot_training - Step 14604: {'lr': 0.00041965211927239644, 'samples': 2804160, 'steps': 14604, 'loss/train': 0.6000882536172867} +01/27/2022 09:35:53 - INFO - codeparrot_training - Step 14605: {'lr': 0.0004196401006777185, 'samples': 2804352, 'steps': 14605, 'loss/train': 0.9335696697235107} +01/27/2022 09:35:56 - INFO - codeparrot_training - Step 14606: {'lr': 0.00041962808135635624, 'samples': 2804544, 'steps': 14606, 'loss/train': 0.36433880031108856} +01/27/2022 09:35:59 - INFO - codeparrot_training - Step 14607: {'lr': 0.00041961606130836105, 'samples': 2804736, 'steps': 14607, 'loss/train': 0.8199672996997833} +01/27/2022 09:36:02 - INFO - codeparrot_training - Step 14608: {'lr': 0.0004196040405337845, 'samples': 2804928, 'steps': 14608, 'loss/train': 0.5985968112945557} +01/27/2022 09:36:05 - INFO - codeparrot_training - Step 14609: {'lr': 0.0004195920190326782, 'samples': 2805120, 'steps': 14609, 'loss/train': 0.2482696697115898} +01/27/2022 09:36:08 - INFO - codeparrot_training - Step 14610: {'lr': 0.0004195799968050935, 'samples': 2805312, 'steps': 14610, 'loss/train': 0.7018756717443466} +01/27/2022 09:36:12 - INFO - codeparrot_training - Step 14611: {'lr': 0.000419567973851082, 'samples': 2805504, 'steps': 14611, 'loss/train': 1.4747478067874908} +01/27/2022 09:36:15 - INFO - codeparrot_training - Step 14612: {'lr': 0.0004195559501706951, 'samples': 2805696, 'steps': 14612, 'loss/train': 0.706634983420372} +01/27/2022 09:36:18 - INFO - codeparrot_training - Step 14613: {'lr': 0.00041954392576398433, 'samples': 2805888, 'steps': 14613, 'loss/train': 0.9508496224880219} +01/27/2022 09:36:22 - INFO - codeparrot_training - Step 14614: {'lr': 0.0004195319006310012, 'samples': 2806080, 'steps': 14614, 'loss/train': 1.0042341649532318} +01/27/2022 09:36:26 - INFO - codeparrot_training - Step 14615: {'lr': 0.0004195198747717973, 'samples': 2806272, 'steps': 14615, 'loss/train': 0.6521348208189011} +01/27/2022 09:36:29 - INFO - codeparrot_training - Step 14616: {'lr': 0.00041950784818642404, 'samples': 2806464, 'steps': 14616, 'loss/train': 0.5851780772209167} +01/27/2022 09:36:32 - INFO - codeparrot_training - Step 14617: {'lr': 0.000419495820874933, 'samples': 2806656, 'steps': 14617, 'loss/train': 0.26271526515483856} +01/27/2022 09:36:35 - INFO - codeparrot_training - Step 14618: {'lr': 0.0004194837928373757, 'samples': 2806848, 'steps': 14618, 'loss/train': 0.5679017454385757} +01/27/2022 09:36:38 - INFO - codeparrot_training - Step 14619: {'lr': 0.0004194717640738036, 'samples': 2807040, 'steps': 14619, 'loss/train': 1.4071779549121857} +01/27/2022 09:36:41 - INFO - codeparrot_training - Step 14620: {'lr': 0.0004194597345842683, 'samples': 2807232, 'steps': 14620, 'loss/train': 1.0447184443473816} +01/27/2022 09:36:44 - INFO - codeparrot_training - Step 14621: {'lr': 0.00041944770436882134, 'samples': 2807424, 'steps': 14621, 'loss/train': 0.8549111187458038} +01/27/2022 09:36:51 - INFO - codeparrot_training - Step 14622: {'lr': 0.00041943567342751423, 'samples': 2807616, 'steps': 14622, 'loss/train': 0.8387998044490814} +01/27/2022 09:36:54 - INFO - codeparrot_training - Step 14623: {'lr': 0.0004194236417603985, 'samples': 2807808, 'steps': 14623, 'loss/train': 0.675030380487442} +01/27/2022 09:36:57 - INFO - codeparrot_training - Step 14624: {'lr': 0.0004194116093675256, 'samples': 2808000, 'steps': 14624, 'loss/train': 0.9604562222957611} +01/27/2022 09:37:00 - INFO - codeparrot_training - Step 14625: {'lr': 0.0004193995762489472, 'samples': 2808192, 'steps': 14625, 'loss/train': 1.0138428211212158} +01/27/2022 09:37:03 - INFO - codeparrot_training - Step 14626: {'lr': 0.0004193875424047148, 'samples': 2808384, 'steps': 14626, 'loss/train': 0.4301336109638214} +01/27/2022 09:37:06 - INFO - codeparrot_training - Step 14627: {'lr': 0.00041937550783488, 'samples': 2808576, 'steps': 14627, 'loss/train': 0.9238438010215759} +01/27/2022 09:37:10 - INFO - codeparrot_training - Step 14628: {'lr': 0.00041936347253949426, 'samples': 2808768, 'steps': 14628, 'loss/train': 0.47743476927280426} +01/27/2022 09:37:13 - INFO - codeparrot_training - Step 14629: {'lr': 0.00041935143651860917, 'samples': 2808960, 'steps': 14629, 'loss/train': 1.0805966556072235} +01/27/2022 09:37:16 - INFO - codeparrot_training - Step 14630: {'lr': 0.0004193393997722764, 'samples': 2809152, 'steps': 14630, 'loss/train': 0.28373701125383377} +01/27/2022 09:37:20 - INFO - codeparrot_training - Step 14631: {'lr': 0.00041932736230054725, 'samples': 2809344, 'steps': 14631, 'loss/train': 0.9410839676856995} +01/27/2022 09:37:24 - INFO - codeparrot_training - Step 14632: {'lr': 0.0004193153241034736, 'samples': 2809536, 'steps': 14632, 'loss/train': 1.0807791352272034} +01/27/2022 09:37:27 - INFO - codeparrot_training - Step 14633: {'lr': 0.00041930328518110675, 'samples': 2809728, 'steps': 14633, 'loss/train': 0.9147649705410004} +01/27/2022 09:37:30 - INFO - codeparrot_training - Step 14634: {'lr': 0.0004192912455334985, 'samples': 2809920, 'steps': 14634, 'loss/train': 1.1978933215141296} +01/27/2022 09:37:33 - INFO - codeparrot_training - Step 14635: {'lr': 0.0004192792051607002, 'samples': 2810112, 'steps': 14635, 'loss/train': 1.463774174451828} +01/27/2022 09:37:36 - INFO - codeparrot_training - Step 14636: {'lr': 0.00041926716406276367, 'samples': 2810304, 'steps': 14636, 'loss/train': 0.7552303969860077} +01/27/2022 09:37:39 - INFO - codeparrot_training - Step 14637: {'lr': 0.0004192551222397402, 'samples': 2810496, 'steps': 14637, 'loss/train': 0.7853117287158966} +01/27/2022 09:37:42 - INFO - codeparrot_training - Step 14638: {'lr': 0.0004192430796916816, 'samples': 2810688, 'steps': 14638, 'loss/train': 0.8517679274082184} +01/27/2022 09:37:46 - INFO - codeparrot_training - Step 14639: {'lr': 0.0004192310364186394, 'samples': 2810880, 'steps': 14639, 'loss/train': 1.0425962805747986} +01/27/2022 09:37:50 - INFO - codeparrot_training - Step 14640: {'lr': 0.0004192189924206652, 'samples': 2811072, 'steps': 14640, 'loss/train': 1.449643850326538} +01/27/2022 09:37:53 - INFO - codeparrot_training - Step 14641: {'lr': 0.0004192069476978105, 'samples': 2811264, 'steps': 14641, 'loss/train': 0.6241938918828964} +01/27/2022 09:37:57 - INFO - codeparrot_training - Step 14642: {'lr': 0.000419194902250127, 'samples': 2811456, 'steps': 14642, 'loss/train': 1.6495562195777893} +01/27/2022 09:38:00 - INFO - codeparrot_training - Step 14643: {'lr': 0.0004191828560776663, 'samples': 2811648, 'steps': 14643, 'loss/train': 0.6785459071397781} +01/27/2022 09:38:03 - INFO - codeparrot_training - Step 14644: {'lr': 0.00041917080918047996, 'samples': 2811840, 'steps': 14644, 'loss/train': 0.6782529205083847} +01/27/2022 09:38:06 - INFO - codeparrot_training - Step 14645: {'lr': 0.00041915876155861954, 'samples': 2812032, 'steps': 14645, 'loss/train': 0.8069485723972321} +01/27/2022 09:38:09 - INFO - codeparrot_training - Step 14646: {'lr': 0.0004191467132121367, 'samples': 2812224, 'steps': 14646, 'loss/train': 0.10885109379887581} +01/27/2022 09:38:12 - INFO - codeparrot_training - Step 14647: {'lr': 0.00041913466414108315, 'samples': 2812416, 'steps': 14647, 'loss/train': 0.8022248446941376} +01/27/2022 09:38:15 - INFO - codeparrot_training - Step 14648: {'lr': 0.0004191226143455103, 'samples': 2812608, 'steps': 14648, 'loss/train': 1.6213234663009644} +01/27/2022 09:38:22 - INFO - codeparrot_training - Step 14649: {'lr': 0.00041911056382546997, 'samples': 2812800, 'steps': 14649, 'loss/train': 0.9419182240962982} +01/27/2022 09:38:25 - INFO - codeparrot_training - Step 14650: {'lr': 0.00041909851258101357, 'samples': 2812992, 'steps': 14650, 'loss/train': 1.0607759356498718} +01/27/2022 09:38:28 - INFO - codeparrot_training - Step 14651: {'lr': 0.0004190864606121929, 'samples': 2813184, 'steps': 14651, 'loss/train': 0.5615985989570618} +01/27/2022 09:38:31 - INFO - codeparrot_training - Step 14652: {'lr': 0.0004190744079190595, 'samples': 2813376, 'steps': 14652, 'loss/train': 0.9481981694698334} +01/27/2022 09:38:34 - INFO - codeparrot_training - Step 14653: {'lr': 0.0004190623545016651, 'samples': 2813568, 'steps': 14653, 'loss/train': 1.1620320081710815} +01/27/2022 09:38:37 - INFO - codeparrot_training - Step 14654: {'lr': 0.00041905030036006106, 'samples': 2813760, 'steps': 14654, 'loss/train': 0.40437206625938416} +01/27/2022 09:38:40 - INFO - codeparrot_training - Step 14655: {'lr': 0.00041903824549429936, 'samples': 2813952, 'steps': 14655, 'loss/train': 0.7149298042058945} +01/27/2022 09:38:44 - INFO - codeparrot_training - Step 14656: {'lr': 0.00041902618990443156, 'samples': 2814144, 'steps': 14656, 'loss/train': 0.4370279610157013} +01/27/2022 09:38:47 - INFO - codeparrot_training - Step 14657: {'lr': 0.0004190141335905091, 'samples': 2814336, 'steps': 14657, 'loss/train': 0.6924601346254349} +01/27/2022 09:38:51 - INFO - codeparrot_training - Step 14658: {'lr': 0.0004190020765525838, 'samples': 2814528, 'steps': 14658, 'loss/train': 1.0857647359371185} +01/27/2022 09:38:55 - INFO - codeparrot_training - Step 14659: {'lr': 0.0004189900187907073, 'samples': 2814720, 'steps': 14659, 'loss/train': 0.9051251113414764} +01/27/2022 09:38:58 - INFO - codeparrot_training - Step 14660: {'lr': 0.0004189779603049312, 'samples': 2814912, 'steps': 14660, 'loss/train': 1.084419697523117} +01/27/2022 09:39:01 - INFO - codeparrot_training - Step 14661: {'lr': 0.00041896590109530713, 'samples': 2815104, 'steps': 14661, 'loss/train': 0.546230748295784} +01/27/2022 09:39:04 - INFO - codeparrot_training - Step 14662: {'lr': 0.00041895384116188685, 'samples': 2815296, 'steps': 14662, 'loss/train': 0.5894006788730621} +01/27/2022 09:39:07 - INFO - codeparrot_training - Step 14663: {'lr': 0.000418941780504722, 'samples': 2815488, 'steps': 14663, 'loss/train': 0.6621887236833572} +01/27/2022 09:39:10 - INFO - codeparrot_training - Step 14664: {'lr': 0.00041892971912386415, 'samples': 2815680, 'steps': 14664, 'loss/train': 0.6637994349002838} +01/27/2022 09:39:13 - INFO - codeparrot_training - Step 14665: {'lr': 0.000418917657019365, 'samples': 2815872, 'steps': 14665, 'loss/train': 0.872938334941864} +01/27/2022 09:39:20 - INFO - codeparrot_training - Step 14666: {'lr': 0.0004189055941912763, 'samples': 2816064, 'steps': 14666, 'loss/train': 0.8275067210197449} +01/27/2022 09:39:23 - INFO - codeparrot_training - Step 14667: {'lr': 0.0004188935306396496, 'samples': 2816256, 'steps': 14667, 'loss/train': 0.8122271597385406} +01/27/2022 09:39:26 - INFO - codeparrot_training - Step 14668: {'lr': 0.00041888146636453674, 'samples': 2816448, 'steps': 14668, 'loss/train': 0.7215463221073151} +01/27/2022 09:39:29 - INFO - codeparrot_training - Step 14669: {'lr': 0.0004188694013659892, 'samples': 2816640, 'steps': 14669, 'loss/train': 1.0571164190769196} +01/27/2022 09:39:32 - INFO - codeparrot_training - Step 14670: {'lr': 0.0004188573356440588, 'samples': 2816832, 'steps': 14670, 'loss/train': 0.8774251341819763} +01/27/2022 09:39:36 - INFO - codeparrot_training - Step 14671: {'lr': 0.0004188452691987973, 'samples': 2817024, 'steps': 14671, 'loss/train': 0.6868015676736832} +01/27/2022 09:39:39 - INFO - codeparrot_training - Step 14672: {'lr': 0.0004188332020302561, 'samples': 2817216, 'steps': 14672, 'loss/train': 0.8803143203258514} +01/27/2022 09:39:42 - INFO - codeparrot_training - Step 14673: {'lr': 0.0004188211341384872, 'samples': 2817408, 'steps': 14673, 'loss/train': 0.8729958236217499} +01/27/2022 09:39:45 - INFO - codeparrot_training - Step 14674: {'lr': 0.0004188090655235421, 'samples': 2817600, 'steps': 14674, 'loss/train': 0.9524068236351013} +01/27/2022 09:39:50 - INFO - codeparrot_training - Step 14675: {'lr': 0.00041879699618547263, 'samples': 2817792, 'steps': 14675, 'loss/train': 0.8621561229228973} +01/27/2022 09:39:53 - INFO - codeparrot_training - Step 14676: {'lr': 0.0004187849261243304, 'samples': 2817984, 'steps': 14676, 'loss/train': 0.5677065700292587} +01/27/2022 09:39:56 - INFO - codeparrot_training - Step 14677: {'lr': 0.0004187728553401671, 'samples': 2818176, 'steps': 14677, 'loss/train': 1.2535160779953003} +01/27/2022 09:39:59 - INFO - codeparrot_training - Step 14678: {'lr': 0.0004187607838330345, 'samples': 2818368, 'steps': 14678, 'loss/train': 0.8035756945610046} +01/27/2022 09:40:02 - INFO - codeparrot_training - Step 14679: {'lr': 0.0004187487116029843, 'samples': 2818560, 'steps': 14679, 'loss/train': 0.6796009540557861} +01/27/2022 09:40:05 - INFO - codeparrot_training - Step 14680: {'lr': 0.0004187366386500683, 'samples': 2818752, 'steps': 14680, 'loss/train': 0.6616907268762589} +01/27/2022 09:40:08 - INFO - codeparrot_training - Step 14681: {'lr': 0.00041872456497433797, 'samples': 2818944, 'steps': 14681, 'loss/train': 0.29991238564252853} +01/27/2022 09:40:11 - INFO - codeparrot_training - Step 14682: {'lr': 0.00041871249057584526, 'samples': 2819136, 'steps': 14682, 'loss/train': 1.0577578246593475} +01/27/2022 09:40:15 - INFO - codeparrot_training - Step 14683: {'lr': 0.00041870041545464176, 'samples': 2819328, 'steps': 14683, 'loss/train': 0.7127761691808701} +01/27/2022 09:40:19 - INFO - codeparrot_training - Step 14684: {'lr': 0.00041868833961077935, 'samples': 2819520, 'steps': 14684, 'loss/train': 0.6910829544067383} +01/27/2022 09:40:22 - INFO - codeparrot_training - Step 14685: {'lr': 0.0004186762630443096, 'samples': 2819712, 'steps': 14685, 'loss/train': 1.0491018891334534} +01/27/2022 09:40:25 - INFO - codeparrot_training - Step 14686: {'lr': 0.0004186641857552842, 'samples': 2819904, 'steps': 14686, 'loss/train': 0.99989452958107} +01/27/2022 09:40:28 - INFO - codeparrot_training - Step 14687: {'lr': 0.0004186521077437551, 'samples': 2820096, 'steps': 14687, 'loss/train': 0.678710550069809} +01/27/2022 09:40:32 - INFO - codeparrot_training - Step 14688: {'lr': 0.00041864002900977393, 'samples': 2820288, 'steps': 14688, 'loss/train': 1.188464641571045} +01/27/2022 09:40:35 - INFO - codeparrot_training - Step 14689: {'lr': 0.0004186279495533923, 'samples': 2820480, 'steps': 14689, 'loss/train': 0.7125510424375534} +01/27/2022 09:40:38 - INFO - codeparrot_training - Step 14690: {'lr': 0.0004186158693746622, 'samples': 2820672, 'steps': 14690, 'loss/train': 1.3235315680503845} +01/27/2022 09:40:41 - INFO - codeparrot_training - Step 14691: {'lr': 0.0004186037884736352, 'samples': 2820864, 'steps': 14691, 'loss/train': 1.0423507690429688} +01/27/2022 09:40:44 - INFO - codeparrot_training - Step 14692: {'lr': 0.0004185917068503632, 'samples': 2821056, 'steps': 14692, 'loss/train': 0.8035167753696442} +01/27/2022 09:40:50 - INFO - codeparrot_training - Step 14693: {'lr': 0.00041857962450489786, 'samples': 2821248, 'steps': 14693, 'loss/train': 0.8102477788925171} +01/27/2022 09:40:54 - INFO - codeparrot_training - Step 14694: {'lr': 0.0004185675414372908, 'samples': 2821440, 'steps': 14694, 'loss/train': 1.450908601284027} +01/27/2022 09:40:57 - INFO - codeparrot_training - Step 14695: {'lr': 0.000418555457647594, 'samples': 2821632, 'steps': 14695, 'loss/train': 0.8337876498699188} +01/27/2022 09:41:00 - INFO - codeparrot_training - Step 14696: {'lr': 0.00041854337313585913, 'samples': 2821824, 'steps': 14696, 'loss/train': 0.9468078017234802} +01/27/2022 09:41:03 - INFO - codeparrot_training - Step 14697: {'lr': 0.00041853128790213804, 'samples': 2822016, 'steps': 14697, 'loss/train': 0.07157945819199085} +01/27/2022 09:41:06 - INFO - codeparrot_training - Step 14698: {'lr': 0.0004185192019464823, 'samples': 2822208, 'steps': 14698, 'loss/train': 0.5361178815364838} +01/27/2022 09:41:09 - INFO - codeparrot_training - Step 14699: {'lr': 0.0004185071152689439, 'samples': 2822400, 'steps': 14699, 'loss/train': 1.1995435953140259} +01/27/2022 09:41:12 - INFO - codeparrot_training - Step 14700: {'lr': 0.0004184950278695745, 'samples': 2822592, 'steps': 14700, 'loss/train': 0.9708455801010132} +01/27/2022 09:41:16 - INFO - codeparrot_training - Step 14701: {'lr': 0.0004184829397484259, 'samples': 2822784, 'steps': 14701, 'loss/train': 1.0767499208450317} +01/27/2022 09:41:20 - INFO - codeparrot_training - Step 14702: {'lr': 0.00041847085090554985, 'samples': 2822976, 'steps': 14702, 'loss/train': 0.7786416113376617} +01/27/2022 09:41:23 - INFO - codeparrot_training - Step 14703: {'lr': 0.00041845876134099825, 'samples': 2823168, 'steps': 14703, 'loss/train': 0.6775613576173782} +01/27/2022 09:41:26 - INFO - codeparrot_training - Step 14704: {'lr': 0.0004184466710548227, 'samples': 2823360, 'steps': 14704, 'loss/train': 1.198413759469986} +01/27/2022 09:41:29 - INFO - codeparrot_training - Step 14705: {'lr': 0.0004184345800470752, 'samples': 2823552, 'steps': 14705, 'loss/train': 0.8965454399585724} +01/27/2022 09:41:33 - INFO - codeparrot_training - Step 14706: {'lr': 0.00041842248831780736, 'samples': 2823744, 'steps': 14706, 'loss/train': 0.8890028893947601} +01/27/2022 09:41:36 - INFO - codeparrot_training - Step 14707: {'lr': 0.0004184103958670712, 'samples': 2823936, 'steps': 14707, 'loss/train': 0.8735033869743347} +01/27/2022 09:41:39 - INFO - codeparrot_training - Step 14708: {'lr': 0.00041839830269491823, 'samples': 2824128, 'steps': 14708, 'loss/train': 0.6840658932924271} +01/27/2022 09:41:42 - INFO - codeparrot_training - Step 14709: {'lr': 0.00041838620880140046, 'samples': 2824320, 'steps': 14709, 'loss/train': 0.36487387865781784} +01/27/2022 09:41:45 - INFO - codeparrot_training - Step 14710: {'lr': 0.00041837411418656965, 'samples': 2824512, 'steps': 14710, 'loss/train': 0.9477590024471283} +01/27/2022 09:41:50 - INFO - codeparrot_training - Step 14711: {'lr': 0.0004183620188504776, 'samples': 2824704, 'steps': 14711, 'loss/train': 0.8802561163902283} +01/27/2022 09:41:53 - INFO - codeparrot_training - Step 14712: {'lr': 0.0004183499227931761, 'samples': 2824896, 'steps': 14712, 'loss/train': 1.2066048681735992} +01/27/2022 09:41:56 - INFO - codeparrot_training - Step 14713: {'lr': 0.00041833782601471704, 'samples': 2825088, 'steps': 14713, 'loss/train': 0.5958120971918106} +01/27/2022 09:41:59 - INFO - codeparrot_training - Step 14714: {'lr': 0.0004183257285151521, 'samples': 2825280, 'steps': 14714, 'loss/train': 0.8856074810028076} +01/27/2022 09:42:02 - INFO - codeparrot_training - Step 14715: {'lr': 0.00041831363029453327, 'samples': 2825472, 'steps': 14715, 'loss/train': 0.7699288129806519} +01/27/2022 09:42:06 - INFO - codeparrot_training - Step 14716: {'lr': 0.0004183015313529123, 'samples': 2825664, 'steps': 14716, 'loss/train': 0.6050240099430084} +01/27/2022 09:42:09 - INFO - codeparrot_training - Step 14717: {'lr': 0.00041828943169034094, 'samples': 2825856, 'steps': 14717, 'loss/train': 1.1114689707756042} +01/27/2022 09:42:12 - INFO - codeparrot_training - Step 14718: {'lr': 0.0004182773313068711, 'samples': 2826048, 'steps': 14718, 'loss/train': 0.6214543282985687} +01/27/2022 09:42:16 - INFO - codeparrot_training - Step 14719: {'lr': 0.00041826523020255463, 'samples': 2826240, 'steps': 14719, 'loss/train': 0.7709759473800659} +01/27/2022 09:42:19 - INFO - codeparrot_training - Step 14720: {'lr': 0.00041825312837744333, 'samples': 2826432, 'steps': 14720, 'loss/train': 0.4294821470975876} +01/27/2022 09:42:22 - INFO - codeparrot_training - Step 14721: {'lr': 0.00041824102583158906, 'samples': 2826624, 'steps': 14721, 'loss/train': 0.45274047553539276} +01/27/2022 09:42:26 - INFO - codeparrot_training - Step 14722: {'lr': 0.0004182289225650437, 'samples': 2826816, 'steps': 14722, 'loss/train': 0.6684993803501129} +01/27/2022 09:42:29 - INFO - codeparrot_training - Step 14723: {'lr': 0.00041821681857785904, 'samples': 2827008, 'steps': 14723, 'loss/train': 0.5814145803451538} +01/27/2022 09:42:32 - INFO - codeparrot_training - Step 14724: {'lr': 0.0004182047138700869, 'samples': 2827200, 'steps': 14724, 'loss/train': 1.171672135591507} +01/27/2022 09:42:35 - INFO - codeparrot_training - Step 14725: {'lr': 0.0004181926084417792, 'samples': 2827392, 'steps': 14725, 'loss/train': 1.204192578792572} +01/27/2022 09:42:38 - INFO - codeparrot_training - Step 14726: {'lr': 0.0004181805022929878, 'samples': 2827584, 'steps': 14726, 'loss/train': 0.13331691920757294} +01/27/2022 09:42:41 - INFO - codeparrot_training - Step 14727: {'lr': 0.0004181683954237645, 'samples': 2827776, 'steps': 14727, 'loss/train': 0.9631253778934479} +01/27/2022 09:42:47 - INFO - codeparrot_training - Step 14728: {'lr': 0.00041815628783416117, 'samples': 2827968, 'steps': 14728, 'loss/train': 0.7458924651145935} +01/27/2022 09:42:51 - INFO - codeparrot_training - Step 14729: {'lr': 0.00041814417952422975, 'samples': 2828160, 'steps': 14729, 'loss/train': 1.2032437026500702} +01/27/2022 09:42:54 - INFO - codeparrot_training - Step 14730: {'lr': 0.000418132070494022, 'samples': 2828352, 'steps': 14730, 'loss/train': 0.7057810574769974} +01/27/2022 09:42:57 - INFO - codeparrot_training - Step 14731: {'lr': 0.00041811996074358993, 'samples': 2828544, 'steps': 14731, 'loss/train': 0.8446931540966034} +01/27/2022 09:43:00 - INFO - codeparrot_training - Step 14732: {'lr': 0.00041810785027298524, 'samples': 2828736, 'steps': 14732, 'loss/train': 1.0381837785243988} +01/27/2022 09:43:03 - INFO - codeparrot_training - Step 14733: {'lr': 0.00041809573908225997, 'samples': 2828928, 'steps': 14733, 'loss/train': 0.8627919852733612} +01/27/2022 09:43:06 - INFO - codeparrot_training - Step 14734: {'lr': 0.00041808362717146594, 'samples': 2829120, 'steps': 14734, 'loss/train': 0.7342803329229355} +01/27/2022 09:43:09 - INFO - codeparrot_training - Step 14735: {'lr': 0.00041807151454065493, 'samples': 2829312, 'steps': 14735, 'loss/train': 0.8722310364246368} +01/27/2022 09:43:13 - INFO - codeparrot_training - Step 14736: {'lr': 0.00041805940118987904, 'samples': 2829504, 'steps': 14736, 'loss/train': 0.9414417743682861} +01/27/2022 09:43:17 - INFO - codeparrot_training - Step 14737: {'lr': 0.0004180472871191899, 'samples': 2829696, 'steps': 14737, 'loss/train': 0.8041000664234161} +01/27/2022 09:43:20 - INFO - codeparrot_training - Step 14738: {'lr': 0.0004180351723286396, 'samples': 2829888, 'steps': 14738, 'loss/train': 1.4101426899433136} +01/27/2022 09:43:23 - INFO - codeparrot_training - Step 14739: {'lr': 0.00041802305681828007, 'samples': 2830080, 'steps': 14739, 'loss/train': 1.277050495147705} +01/27/2022 09:43:26 - INFO - codeparrot_training - Step 14740: {'lr': 0.00041801094058816304, 'samples': 2830272, 'steps': 14740, 'loss/train': 0.9346383512020111} +01/27/2022 09:43:30 - INFO - codeparrot_training - Step 14741: {'lr': 0.0004179988236383405, 'samples': 2830464, 'steps': 14741, 'loss/train': 0.6566660106182098} +01/27/2022 09:43:33 - INFO - codeparrot_training - Step 14742: {'lr': 0.00041798670596886433, 'samples': 2830656, 'steps': 14742, 'loss/train': 0.6905596107244492} +01/27/2022 09:43:36 - INFO - codeparrot_training - Step 14743: {'lr': 0.00041797458757978647, 'samples': 2830848, 'steps': 14743, 'loss/train': 0.6328418701887131} +01/27/2022 09:43:39 - INFO - codeparrot_training - Step 14744: {'lr': 0.0004179624684711588, 'samples': 2831040, 'steps': 14744, 'loss/train': 0.5018843114376068} +01/27/2022 09:43:42 - INFO - codeparrot_training - Step 14745: {'lr': 0.0004179503486430333, 'samples': 2831232, 'steps': 14745, 'loss/train': 1.1646584272384644} +01/27/2022 09:43:48 - INFO - codeparrot_training - Step 14746: {'lr': 0.00041793822809546176, 'samples': 2831424, 'steps': 14746, 'loss/train': 0.34616658836603165} +01/27/2022 09:43:52 - INFO - codeparrot_training - Step 14747: {'lr': 0.0004179261068284963, 'samples': 2831616, 'steps': 14747, 'loss/train': 1.0727170407772064} +01/27/2022 09:43:55 - INFO - codeparrot_training - Step 14748: {'lr': 0.00041791398484218855, 'samples': 2831808, 'steps': 14748, 'loss/train': 0.3609984442591667} +01/27/2022 09:43:58 - INFO - codeparrot_training - Step 14749: {'lr': 0.0004179018621365908, 'samples': 2832000, 'steps': 14749, 'loss/train': 1.0386211574077606} +01/27/2022 09:44:01 - INFO - codeparrot_training - Step 14750: {'lr': 0.00041788973871175465, 'samples': 2832192, 'steps': 14750, 'loss/train': 0.7308953404426575} +01/27/2022 09:44:04 - INFO - codeparrot_training - Step 14751: {'lr': 0.00041787761456773214, 'samples': 2832384, 'steps': 14751, 'loss/train': 1.2049443125724792} +01/27/2022 09:44:07 - INFO - codeparrot_training - Step 14752: {'lr': 0.00041786548970457535, 'samples': 2832576, 'steps': 14752, 'loss/train': 0.7305739223957062} +01/27/2022 09:44:11 - INFO - codeparrot_training - Step 14753: {'lr': 0.000417853364122336, 'samples': 2832768, 'steps': 14753, 'loss/train': 0.5155317038297653} +01/27/2022 09:44:15 - INFO - codeparrot_training - Step 14754: {'lr': 0.0004178412378210662, 'samples': 2832960, 'steps': 14754, 'loss/train': 0.7736903429031372} +01/27/2022 09:44:18 - INFO - codeparrot_training - Step 14755: {'lr': 0.0004178291108008179, 'samples': 2833152, 'steps': 14755, 'loss/train': 0.820461630821228} +01/27/2022 09:44:21 - INFO - codeparrot_training - Step 14756: {'lr': 0.00041781698306164283, 'samples': 2833344, 'steps': 14756, 'loss/train': 0.7758226096630096} +01/27/2022 09:44:24 - INFO - codeparrot_training - Step 14757: {'lr': 0.0004178048546035932, 'samples': 2833536, 'steps': 14757, 'loss/train': 0.9858404695987701} +01/27/2022 09:44:28 - INFO - codeparrot_training - Step 14758: {'lr': 0.00041779272542672086, 'samples': 2833728, 'steps': 14758, 'loss/train': 0.7178333848714828} +01/27/2022 09:44:31 - INFO - codeparrot_training - Step 14759: {'lr': 0.00041778059553107766, 'samples': 2833920, 'steps': 14759, 'loss/train': 0.5144877433776855} +01/27/2022 09:44:34 - INFO - codeparrot_training - Step 14760: {'lr': 0.00041776846491671575, 'samples': 2834112, 'steps': 14760, 'loss/train': 1.1205784678459167} +01/27/2022 09:44:37 - INFO - codeparrot_training - Step 14761: {'lr': 0.000417756333583687, 'samples': 2834304, 'steps': 14761, 'loss/train': 1.210334837436676} +01/27/2022 09:44:40 - INFO - codeparrot_training - Step 14762: {'lr': 0.0004177442015320434, 'samples': 2834496, 'steps': 14762, 'loss/train': 0.7166055142879486} +01/27/2022 09:44:45 - INFO - codeparrot_training - Step 14763: {'lr': 0.0004177320687618369, 'samples': 2834688, 'steps': 14763, 'loss/train': 0.7639804780483246} +01/27/2022 09:44:48 - INFO - codeparrot_training - Step 14764: {'lr': 0.0004177199352731194, 'samples': 2834880, 'steps': 14764, 'loss/train': 0.9898688793182373} +01/27/2022 09:44:51 - INFO - codeparrot_training - Step 14765: {'lr': 0.0004177078010659431, 'samples': 2835072, 'steps': 14765, 'loss/train': 0.8888974785804749} +01/27/2022 09:44:54 - INFO - codeparrot_training - Step 14766: {'lr': 0.0004176956661403597, 'samples': 2835264, 'steps': 14766, 'loss/train': 0.9572798609733582} +01/27/2022 09:44:57 - INFO - codeparrot_training - Step 14767: {'lr': 0.0004176835304964214, 'samples': 2835456, 'steps': 14767, 'loss/train': 0.7814667820930481} +01/27/2022 09:45:00 - INFO - codeparrot_training - Step 14768: {'lr': 0.00041767139413418, 'samples': 2835648, 'steps': 14768, 'loss/train': 0.9931512773036957} +01/27/2022 09:45:03 - INFO - codeparrot_training - Step 14769: {'lr': 0.00041765925705368766, 'samples': 2835840, 'steps': 14769, 'loss/train': 1.0892200469970703} +01/27/2022 09:45:07 - INFO - codeparrot_training - Step 14770: {'lr': 0.00041764711925499633, 'samples': 2836032, 'steps': 14770, 'loss/train': 0.9476015567779541} +01/27/2022 09:45:10 - INFO - codeparrot_training - Step 14771: {'lr': 0.0004176349807381579, 'samples': 2836224, 'steps': 14771, 'loss/train': 0.743904635310173} +01/27/2022 09:45:16 - INFO - codeparrot_training - Step 14772: {'lr': 0.0004176228415032245, 'samples': 2836416, 'steps': 14772, 'loss/train': 1.103442907333374} +01/27/2022 09:45:19 - INFO - codeparrot_training - Step 14773: {'lr': 0.000417610701550248, 'samples': 2836608, 'steps': 14773, 'loss/train': 0.9646082818508148} +01/27/2022 09:45:22 - INFO - codeparrot_training - Step 14774: {'lr': 0.0004175985608792806, 'samples': 2836800, 'steps': 14774, 'loss/train': 2.1541024446487427} +01/27/2022 09:45:25 - INFO - codeparrot_training - Step 14775: {'lr': 0.00041758641949037414, 'samples': 2836992, 'steps': 14775, 'loss/train': 0.9861268401145935} +01/27/2022 09:45:29 - INFO - codeparrot_training - Step 14776: {'lr': 0.00041757427738358066, 'samples': 2837184, 'steps': 14776, 'loss/train': 0.5722086131572723} +01/27/2022 09:45:32 - INFO - codeparrot_training - Step 14777: {'lr': 0.00041756213455895215, 'samples': 2837376, 'steps': 14777, 'loss/train': 0.7353585809469223} +01/27/2022 09:45:35 - INFO - codeparrot_training - Step 14778: {'lr': 0.00041754999101654066, 'samples': 2837568, 'steps': 14778, 'loss/train': 0.8062432408332825} +01/27/2022 09:45:38 - INFO - codeparrot_training - Step 14779: {'lr': 0.0004175378467563983, 'samples': 2837760, 'steps': 14779, 'loss/train': 1.0281025171279907} +01/27/2022 09:45:41 - INFO - codeparrot_training - Step 14780: {'lr': 0.00041752570177857695, 'samples': 2837952, 'steps': 14780, 'loss/train': 0.6528675109148026} +01/27/2022 09:45:46 - INFO - codeparrot_training - Step 14781: {'lr': 0.0004175135560831287, 'samples': 2838144, 'steps': 14781, 'loss/train': 0.7588910758495331} +01/27/2022 09:45:49 - INFO - codeparrot_training - Step 14782: {'lr': 0.00041750140967010554, 'samples': 2838336, 'steps': 14782, 'loss/train': 0.8113317489624023} +01/27/2022 09:45:52 - INFO - codeparrot_training - Step 14783: {'lr': 0.00041748926253955954, 'samples': 2838528, 'steps': 14783, 'loss/train': 0.5977474004030228} +01/27/2022 09:45:55 - INFO - codeparrot_training - Step 14784: {'lr': 0.0004174771146915427, 'samples': 2838720, 'steps': 14784, 'loss/train': 0.8546854555606842} +01/27/2022 09:45:58 - INFO - codeparrot_training - Step 14785: {'lr': 0.00041746496612610705, 'samples': 2838912, 'steps': 14785, 'loss/train': 0.9476097822189331} +01/27/2022 09:46:02 - INFO - codeparrot_training - Step 14786: {'lr': 0.00041745281684330476, 'samples': 2839104, 'steps': 14786, 'loss/train': 0.6566299349069595} +01/27/2022 09:46:05 - INFO - codeparrot_training - Step 14787: {'lr': 0.0004174406668431877, 'samples': 2839296, 'steps': 14787, 'loss/train': 0.8156943619251251} +01/27/2022 09:46:08 - INFO - codeparrot_training - Step 14788: {'lr': 0.000417428516125808, 'samples': 2839488, 'steps': 14788, 'loss/train': 0.9465360045433044} +01/27/2022 09:46:14 - INFO - codeparrot_training - Step 14789: {'lr': 0.0004174163646912178, 'samples': 2839680, 'steps': 14789, 'loss/train': 0.5141174644231796} +01/27/2022 09:46:18 - INFO - codeparrot_training - Step 14790: {'lr': 0.0004174042125394689, 'samples': 2839872, 'steps': 14790, 'loss/train': 0.5665283203125} +01/27/2022 09:46:21 - INFO - codeparrot_training - Step 14791: {'lr': 0.00041739205967061366, 'samples': 2840064, 'steps': 14791, 'loss/train': 0.8038610816001892} +01/27/2022 09:46:24 - INFO - codeparrot_training - Step 14792: {'lr': 0.0004173799060847039, 'samples': 2840256, 'steps': 14792, 'loss/train': 0.536593034863472} +01/27/2022 09:46:27 - INFO - codeparrot_training - Step 14793: {'lr': 0.00041736775178179174, 'samples': 2840448, 'steps': 14793, 'loss/train': 1.1547671556472778} +01/27/2022 09:46:30 - INFO - codeparrot_training - Step 14794: {'lr': 0.0004173555967619294, 'samples': 2840640, 'steps': 14794, 'loss/train': 0.767947643995285} +01/27/2022 09:46:33 - INFO - codeparrot_training - Step 14795: {'lr': 0.00041734344102516873, 'samples': 2840832, 'steps': 14795, 'loss/train': 0.9453931152820587} +01/27/2022 09:46:37 - INFO - codeparrot_training - Step 14796: {'lr': 0.0004173312845715619, 'samples': 2841024, 'steps': 14796, 'loss/train': 0.7022225707769394} +01/27/2022 09:46:40 - INFO - codeparrot_training - Step 14797: {'lr': 0.000417319127401161, 'samples': 2841216, 'steps': 14797, 'loss/train': 0.8349330425262451} +01/27/2022 09:46:44 - INFO - codeparrot_training - Step 14798: {'lr': 0.00041730696951401816, 'samples': 2841408, 'steps': 14798, 'loss/train': 0.9120772182941437} +01/27/2022 09:46:47 - INFO - codeparrot_training - Step 14799: {'lr': 0.00041729481091018527, 'samples': 2841600, 'steps': 14799, 'loss/train': 0.5320842862129211} +01/27/2022 09:46:51 - INFO - codeparrot_training - Step 14800: {'lr': 0.0004172826515897146, 'samples': 2841792, 'steps': 14800, 'loss/train': 0.9921875596046448} +01/27/2022 09:46:54 - INFO - codeparrot_training - Step 14801: {'lr': 0.0004172704915526581, 'samples': 2841984, 'steps': 14801, 'loss/train': 0.7252427190542221} +01/27/2022 09:46:57 - INFO - codeparrot_training - Step 14802: {'lr': 0.000417258330799068, 'samples': 2842176, 'steps': 14802, 'loss/train': 0.9280117750167847} +01/27/2022 09:47:00 - INFO - codeparrot_training - Step 14803: {'lr': 0.00041724616932899627, 'samples': 2842368, 'steps': 14803, 'loss/train': 0.9856557548046112} +01/27/2022 09:47:03 - INFO - codeparrot_training - Step 14804: {'lr': 0.0004172340071424951, 'samples': 2842560, 'steps': 14804, 'loss/train': 1.3009657859802246} +01/27/2022 09:47:06 - INFO - codeparrot_training - Step 14805: {'lr': 0.0004172218442396165, 'samples': 2842752, 'steps': 14805, 'loss/train': 0.722277045249939} +01/27/2022 09:47:11 - INFO - codeparrot_training - Step 14806: {'lr': 0.00041720968062041266, 'samples': 2842944, 'steps': 14806, 'loss/train': 0.9851036667823792} +01/27/2022 09:47:14 - INFO - codeparrot_training - Step 14807: {'lr': 0.0004171975162849356, 'samples': 2843136, 'steps': 14807, 'loss/train': 0.8300107419490814} +01/27/2022 09:47:17 - INFO - codeparrot_training - Step 14808: {'lr': 0.0004171853512332375, 'samples': 2843328, 'steps': 14808, 'loss/train': 1.1342875063419342} +01/27/2022 09:47:20 - INFO - codeparrot_training - Step 14809: {'lr': 0.00041717318546537045, 'samples': 2843520, 'steps': 14809, 'loss/train': 1.038821965456009} +01/27/2022 09:47:23 - INFO - codeparrot_training - Step 14810: {'lr': 0.0004171610189813866, 'samples': 2843712, 'steps': 14810, 'loss/train': 0.6825499534606934} +01/27/2022 09:47:27 - INFO - codeparrot_training - Step 14811: {'lr': 0.000417148851781338, 'samples': 2843904, 'steps': 14811, 'loss/train': 0.9285024404525757} +01/27/2022 09:47:30 - INFO - codeparrot_training - Step 14812: {'lr': 0.0004171366838652767, 'samples': 2844096, 'steps': 14812, 'loss/train': 0.9844288229942322} +01/27/2022 09:47:33 - INFO - codeparrot_training - Step 14813: {'lr': 0.000417124515233255, 'samples': 2844288, 'steps': 14813, 'loss/train': 1.006439745426178} +01/27/2022 09:47:36 - INFO - codeparrot_training - Step 14814: {'lr': 0.00041711234588532497, 'samples': 2844480, 'steps': 14814, 'loss/train': 0.7405475825071335} +01/27/2022 09:47:39 - INFO - codeparrot_training - Step 14815: {'lr': 0.0004171001758215387, 'samples': 2844672, 'steps': 14815, 'loss/train': 0.462926521897316} +01/27/2022 09:47:44 - INFO - codeparrot_training - Step 14816: {'lr': 0.0004170880050419483, 'samples': 2844864, 'steps': 14816, 'loss/train': 1.0326532423496246} +01/27/2022 09:47:47 - INFO - codeparrot_training - Step 14817: {'lr': 0.00041707583354660597, 'samples': 2845056, 'steps': 14817, 'loss/train': 0.8180705308914185} +01/27/2022 09:47:50 - INFO - codeparrot_training - Step 14818: {'lr': 0.0004170636613355638, 'samples': 2845248, 'steps': 14818, 'loss/train': 0.6342102885246277} +01/27/2022 09:47:53 - INFO - codeparrot_training - Step 14819: {'lr': 0.000417051488408874, 'samples': 2845440, 'steps': 14819, 'loss/train': 0.9276755154132843} +01/27/2022 09:47:56 - INFO - codeparrot_training - Step 14820: {'lr': 0.00041703931476658857, 'samples': 2845632, 'steps': 14820, 'loss/train': 0.7134802043437958} +01/27/2022 09:48:00 - INFO - codeparrot_training - Step 14821: {'lr': 0.0004170271404087598, 'samples': 2845824, 'steps': 14821, 'loss/train': 0.7472375929355621} +01/27/2022 09:48:03 - INFO - codeparrot_training - Step 14822: {'lr': 0.0004170149653354398, 'samples': 2846016, 'steps': 14822, 'loss/train': 0.6008272916078568} +01/27/2022 09:48:06 - INFO - codeparrot_training - Step 14823: {'lr': 0.0004170027895466807, 'samples': 2846208, 'steps': 14823, 'loss/train': 1.7794066071510315} +01/27/2022 09:48:12 - INFO - codeparrot_training - Step 14824: {'lr': 0.00041699061304253476, 'samples': 2846400, 'steps': 14824, 'loss/train': 1.2936800122261047} +01/27/2022 09:48:15 - INFO - codeparrot_training - Step 14825: {'lr': 0.00041697843582305406, 'samples': 2846592, 'steps': 14825, 'loss/train': 0.36899325996637344} +01/27/2022 09:48:18 - INFO - codeparrot_training - Step 14826: {'lr': 0.0004169662578882907, 'samples': 2846784, 'steps': 14826, 'loss/train': 1.0826586484909058} +01/27/2022 09:48:22 - INFO - codeparrot_training - Step 14827: {'lr': 0.0004169540792382969, 'samples': 2846976, 'steps': 14827, 'loss/train': 0.28187743574380875} +01/27/2022 09:48:25 - INFO - codeparrot_training - Step 14828: {'lr': 0.0004169418998731249, 'samples': 2847168, 'steps': 14828, 'loss/train': 0.7971208691596985} +01/27/2022 09:48:28 - INFO - codeparrot_training - Step 14829: {'lr': 0.0004169297197928268, 'samples': 2847360, 'steps': 14829, 'loss/train': 1.0158546566963196} +01/27/2022 09:48:31 - INFO - codeparrot_training - Step 14830: {'lr': 0.0004169175389974548, 'samples': 2847552, 'steps': 14830, 'loss/train': 1.2617656588554382} +01/27/2022 09:48:34 - INFO - codeparrot_training - Step 14831: {'lr': 0.0004169053574870609, 'samples': 2847744, 'steps': 14831, 'loss/train': 0.5128538310527802} +01/27/2022 09:48:37 - INFO - codeparrot_training - Step 14832: {'lr': 0.0004168931752616977, 'samples': 2847936, 'steps': 14832, 'loss/train': 1.1525744497776031} +01/27/2022 09:48:40 - INFO - codeparrot_training - Step 14833: {'lr': 0.00041688099232141694, 'samples': 2848128, 'steps': 14833, 'loss/train': 0.6738033592700958} +01/27/2022 09:48:45 - INFO - codeparrot_training - Step 14834: {'lr': 0.0004168688086662711, 'samples': 2848320, 'steps': 14834, 'loss/train': 1.0554452240467072} +01/27/2022 09:48:48 - INFO - codeparrot_training - Step 14835: {'lr': 0.0004168566242963122, 'samples': 2848512, 'steps': 14835, 'loss/train': 1.218626081943512} +01/27/2022 09:48:52 - INFO - codeparrot_training - Step 14836: {'lr': 0.00041684443921159253, 'samples': 2848704, 'steps': 14836, 'loss/train': 0.762505978345871} +01/27/2022 09:48:55 - INFO - codeparrot_training - Step 14837: {'lr': 0.00041683225341216426, 'samples': 2848896, 'steps': 14837, 'loss/train': 0.8371227085590363} +01/27/2022 09:48:58 - INFO - codeparrot_training - Step 14838: {'lr': 0.0004168200668980796, 'samples': 2849088, 'steps': 14838, 'loss/train': 0.9246532917022705} +01/27/2022 09:49:01 - INFO - codeparrot_training - Step 14839: {'lr': 0.0004168078796693908, 'samples': 2849280, 'steps': 14839, 'loss/train': 0.4371764212846756} +01/27/2022 09:49:04 - INFO - codeparrot_training - Step 14840: {'lr': 0.00041679569172614996, 'samples': 2849472, 'steps': 14840, 'loss/train': 1.0923790633678436} +01/27/2022 09:49:07 - INFO - codeparrot_training - Step 14841: {'lr': 0.0004167835030684093, 'samples': 2849664, 'steps': 14841, 'loss/train': 0.37112393975257874} +01/27/2022 09:49:12 - INFO - codeparrot_training - Step 14842: {'lr': 0.0004167713136962211, 'samples': 2849856, 'steps': 14842, 'loss/train': 1.2079759240150452} +01/27/2022 09:49:15 - INFO - codeparrot_training - Step 14843: {'lr': 0.00041675912360963766, 'samples': 2850048, 'steps': 14843, 'loss/train': 0.8484532535076141} +01/27/2022 09:49:18 - INFO - codeparrot_training - Step 14844: {'lr': 0.0004167469328087109, 'samples': 2850240, 'steps': 14844, 'loss/train': 0.9894675314426422} +01/27/2022 09:49:21 - INFO - codeparrot_training - Step 14845: {'lr': 0.0004167347412934933, 'samples': 2850432, 'steps': 14845, 'loss/train': 0.9104496538639069} +01/27/2022 09:49:24 - INFO - codeparrot_training - Step 14846: {'lr': 0.00041672254906403703, 'samples': 2850624, 'steps': 14846, 'loss/train': 0.9841272532939911} +01/27/2022 09:49:27 - INFO - codeparrot_training - Step 14847: {'lr': 0.00041671035612039434, 'samples': 2850816, 'steps': 14847, 'loss/train': 0.902584433555603} +01/27/2022 09:49:31 - INFO - codeparrot_training - Step 14848: {'lr': 0.0004166981624626174, 'samples': 2851008, 'steps': 14848, 'loss/train': 0.606144055724144} +01/27/2022 09:49:34 - INFO - codeparrot_training - Step 14849: {'lr': 0.00041668596809075835, 'samples': 2851200, 'steps': 14849, 'loss/train': 0.7090845108032227} +01/27/2022 09:49:37 - INFO - codeparrot_training - Step 14850: {'lr': 0.0004166737730048697, 'samples': 2851392, 'steps': 14850, 'loss/train': 0.3808351010084152} +01/27/2022 09:49:43 - INFO - codeparrot_training - Step 14851: {'lr': 0.00041666157720500344, 'samples': 2851584, 'steps': 14851, 'loss/train': 1.139484018087387} +01/27/2022 09:49:46 - INFO - codeparrot_training - Step 14852: {'lr': 0.00041664938069121195, 'samples': 2851776, 'steps': 14852, 'loss/train': 0.448562353849411} +01/27/2022 09:49:49 - INFO - codeparrot_training - Step 14853: {'lr': 0.0004166371834635474, 'samples': 2851968, 'steps': 14853, 'loss/train': 1.376759111881256} +01/27/2022 09:49:52 - INFO - codeparrot_training - Step 14854: {'lr': 0.00041662498552206206, 'samples': 2852160, 'steps': 14854, 'loss/train': 0.7254250198602676} +01/27/2022 09:49:56 - INFO - codeparrot_training - Step 14855: {'lr': 0.00041661278686680827, 'samples': 2852352, 'steps': 14855, 'loss/train': 1.3295163810253143} +01/27/2022 09:49:59 - INFO - codeparrot_training - Step 14856: {'lr': 0.00041660058749783813, 'samples': 2852544, 'steps': 14856, 'loss/train': 0.6458389610052109} +01/27/2022 09:50:02 - INFO - codeparrot_training - Step 14857: {'lr': 0.000416588387415204, 'samples': 2852736, 'steps': 14857, 'loss/train': 1.0190350413322449} +01/27/2022 09:50:05 - INFO - codeparrot_training - Step 14858: {'lr': 0.0004165761866189581, 'samples': 2852928, 'steps': 14858, 'loss/train': 0.6567938178777695} +01/27/2022 09:50:08 - INFO - codeparrot_training - Step 14859: {'lr': 0.00041656398510915273, 'samples': 2853120, 'steps': 14859, 'loss/train': 0.892786055803299} +01/27/2022 09:50:14 - INFO - codeparrot_training - Step 14860: {'lr': 0.00041655178288584006, 'samples': 2853312, 'steps': 14860, 'loss/train': 1.0964846312999725} +01/27/2022 09:50:17 - INFO - codeparrot_training - Step 14861: {'lr': 0.00041653957994907255, 'samples': 2853504, 'steps': 14861, 'loss/train': 0.9217387139797211} +01/27/2022 09:50:20 - INFO - codeparrot_training - Step 14862: {'lr': 0.0004165273762989023, 'samples': 2853696, 'steps': 14862, 'loss/train': 1.7883444428443909} +01/27/2022 09:50:23 - INFO - codeparrot_training - Step 14863: {'lr': 0.0004165151719353817, 'samples': 2853888, 'steps': 14863, 'loss/train': 0.7755110263824463} +01/27/2022 09:50:26 - INFO - codeparrot_training - Step 14864: {'lr': 0.0004165029668585629, 'samples': 2854080, 'steps': 14864, 'loss/train': 0.7723269760608673} +01/27/2022 09:50:29 - INFO - codeparrot_training - Step 14865: {'lr': 0.00041649076106849836, 'samples': 2854272, 'steps': 14865, 'loss/train': 1.0524187982082367} +01/27/2022 09:50:33 - INFO - codeparrot_training - Step 14866: {'lr': 0.0004164785545652402, 'samples': 2854464, 'steps': 14866, 'loss/train': 1.0664228796958923} +01/27/2022 09:50:36 - INFO - codeparrot_training - Step 14867: {'lr': 0.0004164663473488408, 'samples': 2854656, 'steps': 14867, 'loss/train': 0.8680561780929565} +01/27/2022 09:50:39 - INFO - codeparrot_training - Step 14868: {'lr': 0.0004164541394193524, 'samples': 2854848, 'steps': 14868, 'loss/train': 0.8964868783950806} +01/27/2022 09:50:45 - INFO - codeparrot_training - Step 14869: {'lr': 0.00041644193077682734, 'samples': 2855040, 'steps': 14869, 'loss/train': 1.0786835253238678} +01/27/2022 09:50:48 - INFO - codeparrot_training - Step 14870: {'lr': 0.0004164297214213179, 'samples': 2855232, 'steps': 14870, 'loss/train': 0.6810383051633835} +01/27/2022 09:50:51 - INFO - codeparrot_training - Step 14871: {'lr': 0.0004164175113528763, 'samples': 2855424, 'steps': 14871, 'loss/train': 1.0401233732700348} +01/27/2022 09:50:54 - INFO - codeparrot_training - Step 14872: {'lr': 0.000416405300571555, 'samples': 2855616, 'steps': 14872, 'loss/train': 0.15714527294039726} +01/27/2022 09:50:58 - INFO - codeparrot_training - Step 14873: {'lr': 0.00041639308907740624, 'samples': 2855808, 'steps': 14873, 'loss/train': 1.154172420501709} +01/27/2022 09:51:01 - INFO - codeparrot_training - Step 14874: {'lr': 0.0004163808768704823, 'samples': 2856000, 'steps': 14874, 'loss/train': 0.67677041888237} +01/27/2022 09:51:04 - INFO - codeparrot_training - Step 14875: {'lr': 0.0004163686639508356, 'samples': 2856192, 'steps': 14875, 'loss/train': 1.0373145639896393} +01/27/2022 09:51:07 - INFO - codeparrot_training - Step 14876: {'lr': 0.00041635645031851826, 'samples': 2856384, 'steps': 14876, 'loss/train': 0.9687960147857666} +01/27/2022 09:51:10 - INFO - codeparrot_training - Step 14877: {'lr': 0.0004163442359735827, 'samples': 2856576, 'steps': 14877, 'loss/train': 0.7696496844291687} +01/27/2022 09:51:15 - INFO - codeparrot_training - Step 14878: {'lr': 0.00041633202091608136, 'samples': 2856768, 'steps': 14878, 'loss/train': 0.847482293844223} +01/27/2022 09:51:18 - INFO - codeparrot_training - Step 14879: {'lr': 0.00041631980514606636, 'samples': 2856960, 'steps': 14879, 'loss/train': 0.6754580587148666} +01/27/2022 09:51:21 - INFO - codeparrot_training - Step 14880: {'lr': 0.0004163075886635902, 'samples': 2857152, 'steps': 14880, 'loss/train': 0.7555694282054901} +01/27/2022 09:51:24 - INFO - codeparrot_training - Step 14881: {'lr': 0.0004162953714687051, 'samples': 2857344, 'steps': 14881, 'loss/train': 0.2570982053875923} +01/27/2022 09:51:27 - INFO - codeparrot_training - Step 14882: {'lr': 0.0004162831535614635, 'samples': 2857536, 'steps': 14882, 'loss/train': 1.0315191149711609} +01/27/2022 09:51:31 - INFO - codeparrot_training - Step 14883: {'lr': 0.0004162709349419176, 'samples': 2857728, 'steps': 14883, 'loss/train': 1.1975763738155365} +01/27/2022 09:51:34 - INFO - codeparrot_training - Step 14884: {'lr': 0.0004162587156101198, 'samples': 2857920, 'steps': 14884, 'loss/train': 0.8490925133228302} +01/27/2022 09:51:37 - INFO - codeparrot_training - Step 14885: {'lr': 0.0004162464955661225, 'samples': 2858112, 'steps': 14885, 'loss/train': 0.42097820341587067} +01/27/2022 09:51:40 - INFO - codeparrot_training - Step 14886: {'lr': 0.000416234274809978, 'samples': 2858304, 'steps': 14886, 'loss/train': 0.5582098513841629} +01/27/2022 09:51:44 - INFO - codeparrot_training - Step 14887: {'lr': 0.00041622205334173863, 'samples': 2858496, 'steps': 14887, 'loss/train': 0.7581454217433929} +01/27/2022 09:51:48 - INFO - codeparrot_training - Step 14888: {'lr': 0.00041620983116145673, 'samples': 2858688, 'steps': 14888, 'loss/train': 1.0134406685829163} +01/27/2022 09:51:51 - INFO - codeparrot_training - Step 14889: {'lr': 0.00041619760826918474, 'samples': 2858880, 'steps': 14889, 'loss/train': 0.46819138526916504} +01/27/2022 09:51:54 - INFO - codeparrot_training - Step 14890: {'lr': 0.00041618538466497496, 'samples': 2859072, 'steps': 14890, 'loss/train': 0.7389763444662094} +01/27/2022 09:51:57 - INFO - codeparrot_training - Step 14891: {'lr': 0.00041617316034887983, 'samples': 2859264, 'steps': 14891, 'loss/train': 0.4043544977903366} +01/27/2022 09:52:00 - INFO - codeparrot_training - Step 14892: {'lr': 0.00041616093532095155, 'samples': 2859456, 'steps': 14892, 'loss/train': 0.6869072467088699} +01/27/2022 09:52:03 - INFO - codeparrot_training - Step 14893: {'lr': 0.00041614870958124264, 'samples': 2859648, 'steps': 14893, 'loss/train': 0.6493963748216629} +01/27/2022 09:52:06 - INFO - codeparrot_training - Step 14894: {'lr': 0.00041613648312980537, 'samples': 2859840, 'steps': 14894, 'loss/train': 1.036612182855606} +01/27/2022 09:52:10 - INFO - codeparrot_training - Step 14895: {'lr': 0.00041612425596669215, 'samples': 2860032, 'steps': 14895, 'loss/train': 2.634717106819153} +01/27/2022 09:52:16 - INFO - codeparrot_training - Step 14896: {'lr': 0.0004161120280919554, 'samples': 2860224, 'steps': 14896, 'loss/train': 1.0464337170124054} +01/27/2022 09:52:19 - INFO - codeparrot_training - Step 14897: {'lr': 0.00041609979950564747, 'samples': 2860416, 'steps': 14897, 'loss/train': 0.9059052765369415} +01/27/2022 09:52:22 - INFO - codeparrot_training - Step 14898: {'lr': 0.00041608757020782073, 'samples': 2860608, 'steps': 14898, 'loss/train': 0.9642875790596008} +01/27/2022 09:52:26 - INFO - codeparrot_training - Step 14899: {'lr': 0.0004160753401985276, 'samples': 2860800, 'steps': 14899, 'loss/train': 0.6371002346277237} +01/27/2022 09:52:29 - INFO - codeparrot_training - Step 14900: {'lr': 0.00041606310947782046, 'samples': 2860992, 'steps': 14900, 'loss/train': 0.6879485696554184} +01/27/2022 09:52:32 - INFO - codeparrot_training - Step 14901: {'lr': 0.00041605087804575167, 'samples': 2861184, 'steps': 14901, 'loss/train': 0.029866931028664112} +01/27/2022 09:52:35 - INFO - codeparrot_training - Step 14902: {'lr': 0.0004160386459023736, 'samples': 2861376, 'steps': 14902, 'loss/train': 1.1892339885234833} +01/27/2022 09:52:38 - INFO - codeparrot_training - Step 14903: {'lr': 0.00041602641304773876, 'samples': 2861568, 'steps': 14903, 'loss/train': 1.1637123227119446} +01/27/2022 09:52:43 - INFO - codeparrot_training - Step 14904: {'lr': 0.0004160141794818995, 'samples': 2861760, 'steps': 14904, 'loss/train': 1.3617397248744965} +01/27/2022 09:52:46 - INFO - codeparrot_training - Step 14905: {'lr': 0.00041600194520490815, 'samples': 2861952, 'steps': 14905, 'loss/train': 0.6246704310178757} +01/27/2022 09:52:49 - INFO - codeparrot_training - Step 14906: {'lr': 0.0004159897102168172, 'samples': 2862144, 'steps': 14906, 'loss/train': 0.9599403440952301} +01/27/2022 09:52:52 - INFO - codeparrot_training - Step 14907: {'lr': 0.00041597747451767905, 'samples': 2862336, 'steps': 14907, 'loss/train': 1.0182712376117706} +01/27/2022 09:52:56 - INFO - codeparrot_training - Step 14908: {'lr': 0.00041596523810754607, 'samples': 2862528, 'steps': 14908, 'loss/train': 0.6133207082748413} +01/27/2022 09:52:59 - INFO - codeparrot_training - Step 14909: {'lr': 0.0004159530009864707, 'samples': 2862720, 'steps': 14909, 'loss/train': 0.7166661322116852} +01/27/2022 09:53:02 - INFO - codeparrot_training - Step 14910: {'lr': 0.0004159407631545054, 'samples': 2862912, 'steps': 14910, 'loss/train': 1.1071895956993103} +01/27/2022 09:53:05 - INFO - codeparrot_training - Step 14911: {'lr': 0.0004159285246117026, 'samples': 2863104, 'steps': 14911, 'loss/train': 0.5593938678503036} +01/27/2022 09:53:08 - INFO - codeparrot_training - Step 14912: {'lr': 0.00041591628535811464, 'samples': 2863296, 'steps': 14912, 'loss/train': 0.8118923306465149} +01/27/2022 09:53:12 - INFO - codeparrot_training - Step 14913: {'lr': 0.000415904045393794, 'samples': 2863488, 'steps': 14913, 'loss/train': 0.4378426820039749} +01/27/2022 09:53:16 - INFO - codeparrot_training - Step 14914: {'lr': 0.0004158918047187931, 'samples': 2863680, 'steps': 14914, 'loss/train': 0.9494974315166473} +01/27/2022 09:53:19 - INFO - codeparrot_training - Step 14915: {'lr': 0.0004158795633331645, 'samples': 2863872, 'steps': 14915, 'loss/train': 1.633556306362152} +01/27/2022 09:53:22 - INFO - codeparrot_training - Step 14916: {'lr': 0.00041586732123696037, 'samples': 2864064, 'steps': 14916, 'loss/train': 0.8532348275184631} +01/27/2022 09:53:25 - INFO - codeparrot_training - Step 14917: {'lr': 0.0004158550784302334, 'samples': 2864256, 'steps': 14917, 'loss/train': 0.8623372614383698} +01/27/2022 09:53:28 - INFO - codeparrot_training - Step 14918: {'lr': 0.0004158428349130359, 'samples': 2864448, 'steps': 14918, 'loss/train': 0.8054812252521515} +01/27/2022 09:53:31 - INFO - codeparrot_training - Step 14919: {'lr': 0.00041583059068542034, 'samples': 2864640, 'steps': 14919, 'loss/train': 0.5904173254966736} +01/27/2022 09:53:34 - INFO - codeparrot_training - Step 14920: {'lr': 0.0004158183457474392, 'samples': 2864832, 'steps': 14920, 'loss/train': 1.0304043889045715} +01/27/2022 09:53:38 - INFO - codeparrot_training - Step 14921: {'lr': 0.00041580610009914486, 'samples': 2865024, 'steps': 14921, 'loss/train': 0.6133888363838196} +01/27/2022 09:53:44 - INFO - codeparrot_training - Step 14922: {'lr': 0.00041579385374058996, 'samples': 2865216, 'steps': 14922, 'loss/train': 0.8990558981895447} +01/27/2022 09:53:47 - INFO - codeparrot_training - Step 14923: {'lr': 0.00041578160667182676, 'samples': 2865408, 'steps': 14923, 'loss/train': 0.8090297877788544} +01/27/2022 09:53:50 - INFO - codeparrot_training - Step 14924: {'lr': 0.00041576935889290777, 'samples': 2865600, 'steps': 14924, 'loss/train': 1.5105382204055786} +01/27/2022 09:53:53 - INFO - codeparrot_training - Step 14925: {'lr': 0.0004157571104038856, 'samples': 2865792, 'steps': 14925, 'loss/train': 0.7566186189651489} +01/27/2022 09:53:56 - INFO - codeparrot_training - Step 14926: {'lr': 0.00041574486120481255, 'samples': 2865984, 'steps': 14926, 'loss/train': 0.8501961529254913} +01/27/2022 09:54:00 - INFO - codeparrot_training - Step 14927: {'lr': 0.0004157326112957411, 'samples': 2866176, 'steps': 14927, 'loss/train': 1.0049327909946442} +01/27/2022 09:54:03 - INFO - codeparrot_training - Step 14928: {'lr': 0.0004157203606767238, 'samples': 2866368, 'steps': 14928, 'loss/train': 0.6804579198360443} +01/27/2022 09:54:06 - INFO - codeparrot_training - Step 14929: {'lr': 0.0004157081093478131, 'samples': 2866560, 'steps': 14929, 'loss/train': 0.252153642475605} +01/27/2022 09:54:09 - INFO - codeparrot_training - Step 14930: {'lr': 0.00041569585730906147, 'samples': 2866752, 'steps': 14930, 'loss/train': 1.0068927705287933} +01/27/2022 09:54:14 - INFO - codeparrot_training - Step 14931: {'lr': 0.0004156836045605214, 'samples': 2866944, 'steps': 14931, 'loss/train': 0.563710168004036} +01/27/2022 09:54:17 - INFO - codeparrot_training - Step 14932: {'lr': 0.0004156713511022454, 'samples': 2867136, 'steps': 14932, 'loss/train': 0.42748913168907166} +01/27/2022 09:54:20 - INFO - codeparrot_training - Step 14933: {'lr': 0.00041565909693428593, 'samples': 2867328, 'steps': 14933, 'loss/train': 1.3977684080600739} +01/27/2022 09:54:23 - INFO - codeparrot_training - Step 14934: {'lr': 0.00041564684205669546, 'samples': 2867520, 'steps': 14934, 'loss/train': 0.846125453710556} +01/27/2022 09:54:26 - INFO - codeparrot_training - Step 14935: {'lr': 0.00041563458646952655, 'samples': 2867712, 'steps': 14935, 'loss/train': 0.7471296787261963} +01/27/2022 09:54:29 - INFO - codeparrot_training - Step 14936: {'lr': 0.0004156223301728316, 'samples': 2867904, 'steps': 14936, 'loss/train': 0.8648886680603027} +01/27/2022 09:54:32 - INFO - codeparrot_training - Step 14937: {'lr': 0.00041561007316666333, 'samples': 2868096, 'steps': 14937, 'loss/train': 0.28202104568481445} +01/27/2022 09:54:36 - INFO - codeparrot_training - Step 14938: {'lr': 0.00041559781545107393, 'samples': 2868288, 'steps': 14938, 'loss/train': 0.5227208286523819} +01/27/2022 09:54:40 - INFO - codeparrot_training - Step 14939: {'lr': 0.00041558555702611615, 'samples': 2868480, 'steps': 14939, 'loss/train': 0.774116188287735} +01/27/2022 09:54:43 - INFO - codeparrot_training - Step 14940: {'lr': 0.0004155732978918424, 'samples': 2868672, 'steps': 14940, 'loss/train': 1.1576021611690521} +01/27/2022 09:54:46 - INFO - codeparrot_training - Step 14941: {'lr': 0.00041556103804830523, 'samples': 2868864, 'steps': 14941, 'loss/train': 0.8798396587371826} +01/27/2022 09:54:50 - INFO - codeparrot_training - Step 14942: {'lr': 0.0004155487774955572, 'samples': 2869056, 'steps': 14942, 'loss/train': 1.0762514770030975} +01/27/2022 09:54:53 - INFO - codeparrot_training - Step 14943: {'lr': 0.00041553651623365076, 'samples': 2869248, 'steps': 14943, 'loss/train': 1.0704386830329895} +01/27/2022 09:54:56 - INFO - codeparrot_training - Step 14944: {'lr': 0.00041552425426263836, 'samples': 2869440, 'steps': 14944, 'loss/train': 0.028738743625581264} +01/27/2022 09:54:59 - INFO - codeparrot_training - Step 14945: {'lr': 0.00041551199158257264, 'samples': 2869632, 'steps': 14945, 'loss/train': 0.8262295424938202} +01/27/2022 09:55:02 - INFO - codeparrot_training - Step 14946: {'lr': 0.00041549972819350615, 'samples': 2869824, 'steps': 14946, 'loss/train': 1.181008905172348} +01/27/2022 09:55:05 - INFO - codeparrot_training - Step 14947: {'lr': 0.00041548746409549134, 'samples': 2870016, 'steps': 14947, 'loss/train': 1.2898524105548859} +01/27/2022 09:55:10 - INFO - codeparrot_training - Step 14948: {'lr': 0.0004154751992885808, 'samples': 2870208, 'steps': 14948, 'loss/train': 0.9211480915546417} +01/27/2022 09:55:13 - INFO - codeparrot_training - Step 14949: {'lr': 0.0004154629337728271, 'samples': 2870400, 'steps': 14949, 'loss/train': 0.6386914998292923} +01/27/2022 09:55:16 - INFO - codeparrot_training - Step 14950: {'lr': 0.00041545066754828264, 'samples': 2870592, 'steps': 14950, 'loss/train': 0.6085518300533295} +01/27/2022 09:55:19 - INFO - codeparrot_training - Step 14951: {'lr': 0.00041543840061500007, 'samples': 2870784, 'steps': 14951, 'loss/train': 0.8187743425369263} +01/27/2022 09:55:22 - INFO - codeparrot_training - Step 14952: {'lr': 0.000415426132973032, 'samples': 2870976, 'steps': 14952, 'loss/train': 0.9395715594291687} +01/27/2022 09:55:26 - INFO - codeparrot_training - Step 14953: {'lr': 0.0004154138646224308, 'samples': 2871168, 'steps': 14953, 'loss/train': 0.9904250800609589} +01/27/2022 09:55:29 - INFO - codeparrot_training - Step 14954: {'lr': 0.0004154015955632492, 'samples': 2871360, 'steps': 14954, 'loss/train': 0.841172844171524} +01/27/2022 09:55:32 - INFO - codeparrot_training - Step 14955: {'lr': 0.0004153893257955397, 'samples': 2871552, 'steps': 14955, 'loss/train': 0.9992859363555908} +01/27/2022 09:55:35 - INFO - codeparrot_training - Step 14956: {'lr': 0.00041537705531935476, 'samples': 2871744, 'steps': 14956, 'loss/train': 0.5827158540487289} +01/27/2022 09:55:41 - INFO - codeparrot_training - Step 14957: {'lr': 0.0004153647841347471, 'samples': 2871936, 'steps': 14957, 'loss/train': 1.140465795993805} +01/27/2022 09:55:44 - INFO - codeparrot_training - Step 14958: {'lr': 0.0004153525122417692, 'samples': 2872128, 'steps': 14958, 'loss/train': 0.7829126715660095} +01/27/2022 09:55:48 - INFO - codeparrot_training - Step 14959: {'lr': 0.00041534023964047363, 'samples': 2872320, 'steps': 14959, 'loss/train': 1.0058841705322266} +01/27/2022 09:55:51 - INFO - codeparrot_training - Step 14960: {'lr': 0.00041532796633091297, 'samples': 2872512, 'steps': 14960, 'loss/train': 0.5565561354160309} +01/27/2022 09:55:54 - INFO - codeparrot_training - Step 14961: {'lr': 0.0004153156923131398, 'samples': 2872704, 'steps': 14961, 'loss/train': 1.2491376399993896} +01/27/2022 09:55:57 - INFO - codeparrot_training - Step 14962: {'lr': 0.0004153034175872067, 'samples': 2872896, 'steps': 14962, 'loss/train': 1.0159537196159363} +01/27/2022 09:56:00 - INFO - codeparrot_training - Step 14963: {'lr': 0.00041529114215316633, 'samples': 2873088, 'steps': 14963, 'loss/train': 0.38084784150123596} +01/27/2022 09:56:03 - INFO - codeparrot_training - Step 14964: {'lr': 0.0004152788660110711, 'samples': 2873280, 'steps': 14964, 'loss/train': 0.6386151909828186} +01/27/2022 09:56:06 - INFO - codeparrot_training - Step 14965: {'lr': 0.0004152665891609737, 'samples': 2873472, 'steps': 14965, 'loss/train': 0.6162815541028976} +01/27/2022 09:56:11 - INFO - codeparrot_training - Step 14966: {'lr': 0.0004152543116029267, 'samples': 2873664, 'steps': 14966, 'loss/train': 0.7855096757411957} +01/27/2022 09:56:14 - INFO - codeparrot_training - Step 14967: {'lr': 0.0004152420333369827, 'samples': 2873856, 'steps': 14967, 'loss/train': 1.015951931476593} +01/27/2022 09:56:17 - INFO - codeparrot_training - Step 14968: {'lr': 0.00041522975436319445, 'samples': 2874048, 'steps': 14968, 'loss/train': 0.8404653668403625} +01/27/2022 09:56:21 - INFO - codeparrot_training - Step 14969: {'lr': 0.00041521747468161417, 'samples': 2874240, 'steps': 14969, 'loss/train': 0.6196260452270508} +01/27/2022 09:56:24 - INFO - codeparrot_training - Step 14970: {'lr': 0.00041520519429229485, 'samples': 2874432, 'steps': 14970, 'loss/train': 1.113535076379776} +01/27/2022 09:56:27 - INFO - codeparrot_training - Step 14971: {'lr': 0.00041519291319528886, 'samples': 2874624, 'steps': 14971, 'loss/train': 0.39278528094291687} +01/27/2022 09:56:30 - INFO - codeparrot_training - Step 14972: {'lr': 0.00041518063139064893, 'samples': 2874816, 'steps': 14972, 'loss/train': 0.43085020780563354} +01/27/2022 09:56:33 - INFO - codeparrot_training - Step 14973: {'lr': 0.0004151683488784276, 'samples': 2875008, 'steps': 14973, 'loss/train': 0.7466789782047272} +01/27/2022 09:56:36 - INFO - codeparrot_training - Step 14974: {'lr': 0.00041515606565867746, 'samples': 2875200, 'steps': 14974, 'loss/train': 0.965239405632019} +01/27/2022 09:56:43 - INFO - codeparrot_training - Step 14975: {'lr': 0.0004151437817314513, 'samples': 2875392, 'steps': 14975, 'loss/train': 0.8094145953655243} +01/27/2022 09:56:46 - INFO - codeparrot_training - Step 14976: {'lr': 0.00041513149709680155, 'samples': 2875584, 'steps': 14976, 'loss/train': 0.8900536000728607} +01/27/2022 09:56:49 - INFO - codeparrot_training - Step 14977: {'lr': 0.00041511921175478085, 'samples': 2875776, 'steps': 14977, 'loss/train': 0.42175009846687317} +01/27/2022 09:56:52 - INFO - codeparrot_training - Step 14978: {'lr': 0.0004151069257054419, 'samples': 2875968, 'steps': 14978, 'loss/train': 0.8767218589782715} +01/27/2022 09:56:55 - INFO - codeparrot_training - Step 14979: {'lr': 0.0004150946389488374, 'samples': 2876160, 'steps': 14979, 'loss/train': 1.1750462651252747} +01/27/2022 09:56:58 - INFO - codeparrot_training - Step 14980: {'lr': 0.0004150823514850198, 'samples': 2876352, 'steps': 14980, 'loss/train': 0.2686251327395439} +01/27/2022 09:57:01 - INFO - codeparrot_training - Step 14981: {'lr': 0.00041507006331404186, 'samples': 2876544, 'steps': 14981, 'loss/train': 0.8051825165748596} +01/27/2022 09:57:05 - INFO - codeparrot_training - Step 14982: {'lr': 0.00041505777443595615, 'samples': 2876736, 'steps': 14982, 'loss/train': 0.9834910333156586} +01/27/2022 09:57:09 - INFO - codeparrot_training - Step 14983: {'lr': 0.0004150454848508154, 'samples': 2876928, 'steps': 14983, 'loss/train': 0.6123595833778381} +01/27/2022 09:57:12 - INFO - codeparrot_training - Step 14984: {'lr': 0.00041503319455867216, 'samples': 2877120, 'steps': 14984, 'loss/train': 0.811616063117981} +01/27/2022 09:57:15 - INFO - codeparrot_training - Step 14985: {'lr': 0.0004150209035595791, 'samples': 2877312, 'steps': 14985, 'loss/train': 0.5834321826696396} +01/27/2022 09:57:18 - INFO - codeparrot_training - Step 14986: {'lr': 0.000415008611853589, 'samples': 2877504, 'steps': 14986, 'loss/train': 0.7463065981864929} +01/27/2022 09:57:22 - INFO - codeparrot_training - Step 14987: {'lr': 0.0004149963194407543, 'samples': 2877696, 'steps': 14987, 'loss/train': 0.8775812387466431} +01/27/2022 09:57:25 - INFO - codeparrot_training - Step 14988: {'lr': 0.00041498402632112776, 'samples': 2877888, 'steps': 14988, 'loss/train': 0.9469330608844757} +01/27/2022 09:57:28 - INFO - codeparrot_training - Step 14989: {'lr': 0.00041497173249476204, 'samples': 2878080, 'steps': 14989, 'loss/train': 1.1701518595218658} +01/27/2022 09:57:31 - INFO - codeparrot_training - Step 14990: {'lr': 0.0004149594379617099, 'samples': 2878272, 'steps': 14990, 'loss/train': 1.138960987329483} +01/27/2022 09:57:34 - INFO - codeparrot_training - Step 14991: {'lr': 0.00041494714272202385, 'samples': 2878464, 'steps': 14991, 'loss/train': 0.9694905281066895} +01/27/2022 09:57:39 - INFO - codeparrot_training - Step 14992: {'lr': 0.00041493484677575655, 'samples': 2878656, 'steps': 14992, 'loss/train': 0.886474996805191} +01/27/2022 09:57:42 - INFO - codeparrot_training - Step 14993: {'lr': 0.00041492255012296077, 'samples': 2878848, 'steps': 14993, 'loss/train': 1.0491735935211182} +01/27/2022 09:57:45 - INFO - codeparrot_training - Step 14994: {'lr': 0.0004149102527636892, 'samples': 2879040, 'steps': 14994, 'loss/train': 0.7877329587936401} +01/27/2022 09:57:48 - INFO - codeparrot_training - Step 14995: {'lr': 0.0004148979546979944, 'samples': 2879232, 'steps': 14995, 'loss/train': 0.7126472890377045} +01/27/2022 09:57:51 - INFO - codeparrot_training - Step 14996: {'lr': 0.00041488565592592917, 'samples': 2879424, 'steps': 14996, 'loss/train': 0.5102161020040512} +01/27/2022 09:57:54 - INFO - codeparrot_training - Step 14997: {'lr': 0.0004148733564475462, 'samples': 2879616, 'steps': 14997, 'loss/train': 0.8605283796787262} +01/27/2022 09:57:58 - INFO - codeparrot_training - Step 14998: {'lr': 0.000414861056262898, 'samples': 2879808, 'steps': 14998, 'loss/train': 0.6425975114107132} +01/27/2022 09:58:01 - INFO - codeparrot_training - Step 14999: {'lr': 0.0004148487553720375, 'samples': 2880000, 'steps': 14999, 'loss/train': 0.6975443065166473} +01/27/2022 09:58:04 - INFO - codeparrot_training - Step 15000: {'lr': 0.0004148364537750172, 'samples': 2880192, 'steps': 15000, 'loss/train': 1.240765392780304} +01/27/2022 09:58:10 - INFO - codeparrot_training - Step 15001: {'lr': 0.0004148241514718899, 'samples': 2880384, 'steps': 15001, 'loss/train': 0.9081425964832306} +01/27/2022 09:58:13 - INFO - codeparrot_training - Step 15002: {'lr': 0.00041481184846270836, 'samples': 2880576, 'steps': 15002, 'loss/train': 0.9218344688415527} +01/27/2022 09:58:16 - INFO - codeparrot_training - Step 15003: {'lr': 0.00041479954474752507, 'samples': 2880768, 'steps': 15003, 'loss/train': 0.5083440989255905} +01/27/2022 09:58:20 - INFO - codeparrot_training - Step 15004: {'lr': 0.0004147872403263929, 'samples': 2880960, 'steps': 15004, 'loss/train': 0.9584490358829498} +01/27/2022 09:58:23 - INFO - codeparrot_training - Step 15005: {'lr': 0.0004147749351993645, 'samples': 2881152, 'steps': 15005, 'loss/train': 0.642764613032341} +01/27/2022 09:58:26 - INFO - codeparrot_training - Step 15006: {'lr': 0.0004147626293664926, 'samples': 2881344, 'steps': 15006, 'loss/train': 0.6055668443441391} +01/27/2022 09:58:29 - INFO - codeparrot_training - Step 15007: {'lr': 0.00041475032282783, 'samples': 2881536, 'steps': 15007, 'loss/train': 0.7532553970813751} +01/27/2022 09:58:32 - INFO - codeparrot_training - Step 15008: {'lr': 0.0004147380155834293, 'samples': 2881728, 'steps': 15008, 'loss/train': 1.5778334140777588} +01/27/2022 09:58:35 - INFO - codeparrot_training - Step 15009: {'lr': 0.00041472570763334316, 'samples': 2881920, 'steps': 15009, 'loss/train': 1.0040399730205536} +01/27/2022 09:58:40 - INFO - codeparrot_training - Step 15010: {'lr': 0.00041471339897762447, 'samples': 2882112, 'steps': 15010, 'loss/train': 0.8009811043739319} +01/27/2022 09:58:43 - INFO - codeparrot_training - Step 15011: {'lr': 0.0004147010896163259, 'samples': 2882304, 'steps': 15011, 'loss/train': 1.0453627109527588} +01/27/2022 09:58:46 - INFO - codeparrot_training - Step 15012: {'lr': 0.00041468877954950006, 'samples': 2882496, 'steps': 15012, 'loss/train': 0.7006141394376755} +01/27/2022 09:58:49 - INFO - codeparrot_training - Step 15013: {'lr': 0.0004146764687771999, 'samples': 2882688, 'steps': 15013, 'loss/train': 0.48668789863586426} +01/27/2022 09:58:52 - INFO - codeparrot_training - Step 15014: {'lr': 0.00041466415729947794, 'samples': 2882880, 'steps': 15014, 'loss/train': 0.9708479046821594} +01/27/2022 09:58:55 - INFO - codeparrot_training - Step 15015: {'lr': 0.0004146518451163871, 'samples': 2883072, 'steps': 15015, 'loss/train': 0.7377860248088837} +01/27/2022 09:58:59 - INFO - codeparrot_training - Step 15016: {'lr': 0.00041463953222798, 'samples': 2883264, 'steps': 15016, 'loss/train': 0.7226182222366333} +01/27/2022 09:59:02 - INFO - codeparrot_training - Step 15017: {'lr': 0.00041462721863430943, 'samples': 2883456, 'steps': 15017, 'loss/train': 0.9016934037208557} +01/27/2022 09:59:05 - INFO - codeparrot_training - Step 15018: {'lr': 0.0004146149043354281, 'samples': 2883648, 'steps': 15018, 'loss/train': 0.9316791594028473} +01/27/2022 09:59:10 - INFO - codeparrot_training - Step 15019: {'lr': 0.0004146025893313888, 'samples': 2883840, 'steps': 15019, 'loss/train': 0.39896075427532196} +01/27/2022 09:59:13 - INFO - codeparrot_training - Step 15020: {'lr': 0.00041459027362224433, 'samples': 2884032, 'steps': 15020, 'loss/train': 0.9398731291294098} +01/27/2022 09:59:16 - INFO - codeparrot_training - Step 15021: {'lr': 0.0004145779572080473, 'samples': 2884224, 'steps': 15021, 'loss/train': 1.0990937054157257} +01/27/2022 09:59:19 - INFO - codeparrot_training - Step 15022: {'lr': 0.0004145656400888506, 'samples': 2884416, 'steps': 15022, 'loss/train': 1.0989840924739838} +01/27/2022 09:59:22 - INFO - codeparrot_training - Step 15023: {'lr': 0.000414553322264707, 'samples': 2884608, 'steps': 15023, 'loss/train': 0.7557978630065918} +01/27/2022 09:59:25 - INFO - codeparrot_training - Step 15024: {'lr': 0.00041454100373566915, 'samples': 2884800, 'steps': 15024, 'loss/train': 0.5843474417924881} +01/27/2022 09:59:28 - INFO - codeparrot_training - Step 15025: {'lr': 0.00041452868450178994, 'samples': 2884992, 'steps': 15025, 'loss/train': 0.46224354207515717} +01/27/2022 09:59:32 - INFO - codeparrot_training - Step 15026: {'lr': 0.00041451636456312207, 'samples': 2885184, 'steps': 15026, 'loss/train': 1.195622831583023} +01/27/2022 09:59:38 - INFO - codeparrot_training - Step 15027: {'lr': 0.0004145040439197183, 'samples': 2885376, 'steps': 15027, 'loss/train': 0.8623473644256592} +01/27/2022 09:59:41 - INFO - codeparrot_training - Step 15028: {'lr': 0.00041449172257163156, 'samples': 2885568, 'steps': 15028, 'loss/train': 0.4661318063735962} +01/27/2022 09:59:44 - INFO - codeparrot_training - Step 15029: {'lr': 0.00041447940051891435, 'samples': 2885760, 'steps': 15029, 'loss/train': 1.0171912908554077} +01/27/2022 09:59:47 - INFO - codeparrot_training - Step 15030: {'lr': 0.00041446707776161975, 'samples': 2885952, 'steps': 15030, 'loss/train': 0.6461917161941528} +01/27/2022 09:59:50 - INFO - codeparrot_training - Step 15031: {'lr': 0.00041445475429980033, 'samples': 2886144, 'steps': 15031, 'loss/train': 0.7885690927505493} +01/27/2022 09:59:54 - INFO - codeparrot_training - Step 15032: {'lr': 0.000414442430133509, 'samples': 2886336, 'steps': 15032, 'loss/train': 0.6364949494600296} +01/27/2022 09:59:57 - INFO - codeparrot_training - Step 15033: {'lr': 0.0004144301052627985, 'samples': 2886528, 'steps': 15033, 'loss/train': 0.48566973209381104} +01/27/2022 10:00:00 - INFO - codeparrot_training - Step 15034: {'lr': 0.00041441777968772165, 'samples': 2886720, 'steps': 15034, 'loss/train': 0.21621376276016235} +01/27/2022 10:00:03 - INFO - codeparrot_training - Step 15035: {'lr': 0.00041440545340833124, 'samples': 2886912, 'steps': 15035, 'loss/train': 2.3699668049812317} +01/27/2022 10:00:08 - INFO - codeparrot_training - Step 15036: {'lr': 0.00041439312642468007, 'samples': 2887104, 'steps': 15036, 'loss/train': 0.9903697371482849} +01/27/2022 10:00:11 - INFO - codeparrot_training - Step 15037: {'lr': 0.000414380798736821, 'samples': 2887296, 'steps': 15037, 'loss/train': 0.6954662650823593} +01/27/2022 10:00:14 - INFO - codeparrot_training - Step 15038: {'lr': 0.0004143684703448067, 'samples': 2887488, 'steps': 15038, 'loss/train': 0.7904753386974335} +01/27/2022 10:00:17 - INFO - codeparrot_training - Step 15039: {'lr': 0.0004143561412486901, 'samples': 2887680, 'steps': 15039, 'loss/train': 0.9077926576137543} +01/27/2022 10:00:20 - INFO - codeparrot_training - Step 15040: {'lr': 0.00041434381144852395, 'samples': 2887872, 'steps': 15040, 'loss/train': 0.7395349591970444} +01/27/2022 10:00:24 - INFO - codeparrot_training - Step 15041: {'lr': 0.00041433148094436115, 'samples': 2888064, 'steps': 15041, 'loss/train': 0.7252790182828903} +01/27/2022 10:00:27 - INFO - codeparrot_training - Step 15042: {'lr': 0.0004143191497362544, 'samples': 2888256, 'steps': 15042, 'loss/train': 0.9122817814350128} +01/27/2022 10:00:30 - INFO - codeparrot_training - Step 15043: {'lr': 0.0004143068178242566, 'samples': 2888448, 'steps': 15043, 'loss/train': 1.2050303220748901} +01/27/2022 10:00:33 - INFO - codeparrot_training - Step 15044: {'lr': 0.00041429448520842064, 'samples': 2888640, 'steps': 15044, 'loss/train': 0.7303286343812943} +01/27/2022 10:00:38 - INFO - codeparrot_training - Step 15045: {'lr': 0.00041428215188879926, 'samples': 2888832, 'steps': 15045, 'loss/train': 0.7626270353794098} +01/27/2022 10:00:41 - INFO - codeparrot_training - Step 15046: {'lr': 0.0004142698178654453, 'samples': 2889024, 'steps': 15046, 'loss/train': 0.60106261074543} +01/27/2022 10:00:44 - INFO - codeparrot_training - Step 15047: {'lr': 0.0004142574831384115, 'samples': 2889216, 'steps': 15047, 'loss/train': 1.0041577219963074} +01/27/2022 10:00:47 - INFO - codeparrot_training - Step 15048: {'lr': 0.0004142451477077509, 'samples': 2889408, 'steps': 15048, 'loss/train': 0.8390726745128632} +01/27/2022 10:00:50 - INFO - codeparrot_training - Step 15049: {'lr': 0.00041423281157351624, 'samples': 2889600, 'steps': 15049, 'loss/train': 1.0486519932746887} +01/27/2022 10:00:53 - INFO - codeparrot_training - Step 15050: {'lr': 0.00041422047473576033, 'samples': 2889792, 'steps': 15050, 'loss/train': 1.1063047349452972} +01/27/2022 10:00:56 - INFO - codeparrot_training - Step 15051: {'lr': 0.0004142081371945361, 'samples': 2889984, 'steps': 15051, 'loss/train': 0.477162167429924} +01/27/2022 10:01:00 - INFO - codeparrot_training - Step 15052: {'lr': 0.00041419579894989633, 'samples': 2890176, 'steps': 15052, 'loss/train': 0.5751157701015472} +01/27/2022 10:01:03 - INFO - codeparrot_training - Step 15053: {'lr': 0.0004141834600018939, 'samples': 2890368, 'steps': 15053, 'loss/train': 0.8466656506061554} +01/27/2022 10:01:09 - INFO - codeparrot_training - Step 15054: {'lr': 0.00041417112035058157, 'samples': 2890560, 'steps': 15054, 'loss/train': 0.935164600610733} +01/27/2022 10:01:12 - INFO - codeparrot_training - Step 15055: {'lr': 0.00041415877999601236, 'samples': 2890752, 'steps': 15055, 'loss/train': 0.9793701767921448} +01/27/2022 10:01:15 - INFO - codeparrot_training - Step 15056: {'lr': 0.0004141464389382391, 'samples': 2890944, 'steps': 15056, 'loss/train': 0.8182620406150818} +01/27/2022 10:01:18 - INFO - codeparrot_training - Step 15057: {'lr': 0.0004141340971773147, 'samples': 2891136, 'steps': 15057, 'loss/train': 0.9918774962425232} +01/27/2022 10:01:22 - INFO - codeparrot_training - Step 15058: {'lr': 0.00041412175471329174, 'samples': 2891328, 'steps': 15058, 'loss/train': 0.3347039446234703} +01/27/2022 10:01:25 - INFO - codeparrot_training - Step 15059: {'lr': 0.0004141094115462234, 'samples': 2891520, 'steps': 15059, 'loss/train': 0.7202198803424835} +01/27/2022 10:01:28 - INFO - codeparrot_training - Step 15060: {'lr': 0.00041409706767616246, 'samples': 2891712, 'steps': 15060, 'loss/train': 0.8116297423839569} +01/27/2022 10:01:31 - INFO - codeparrot_training - Step 15061: {'lr': 0.0004140847231031618, 'samples': 2891904, 'steps': 15061, 'loss/train': 0.2850605249404907} +01/27/2022 10:01:34 - INFO - codeparrot_training - Step 15062: {'lr': 0.00041407237782727427, 'samples': 2892096, 'steps': 15062, 'loss/train': 0.9285447299480438} +01/27/2022 10:01:39 - INFO - codeparrot_training - Step 15063: {'lr': 0.0004140600318485527, 'samples': 2892288, 'steps': 15063, 'loss/train': 1.611127495765686} +01/27/2022 10:01:42 - INFO - codeparrot_training - Step 15064: {'lr': 0.0004140476851670502, 'samples': 2892480, 'steps': 15064, 'loss/train': 0.6915925741195679} +01/27/2022 10:01:45 - INFO - codeparrot_training - Step 15065: {'lr': 0.00041403533778281934, 'samples': 2892672, 'steps': 15065, 'loss/train': 0.3854217231273651} +01/27/2022 10:01:48 - INFO - codeparrot_training - Step 15066: {'lr': 0.0004140229896959132, 'samples': 2892864, 'steps': 15066, 'loss/train': 0.7096794694662094} +01/27/2022 10:01:51 - INFO - codeparrot_training - Step 15067: {'lr': 0.00041401064090638474, 'samples': 2893056, 'steps': 15067, 'loss/train': 0.3744534105062485} +01/27/2022 10:01:54 - INFO - codeparrot_training - Step 15068: {'lr': 0.0004139982914142868, 'samples': 2893248, 'steps': 15068, 'loss/train': 1.241644710302353} +01/27/2022 10:01:58 - INFO - codeparrot_training - Step 15069: {'lr': 0.00041398594121967215, 'samples': 2893440, 'steps': 15069, 'loss/train': 0.9711930155754089} +01/27/2022 10:02:01 - INFO - codeparrot_training - Step 15070: {'lr': 0.0004139735903225939, 'samples': 2893632, 'steps': 15070, 'loss/train': 1.2107165157794952} +01/27/2022 10:02:04 - INFO - codeparrot_training - Step 15071: {'lr': 0.0004139612387231048, 'samples': 2893824, 'steps': 15071, 'loss/train': 0.13207560405135155} +01/27/2022 10:02:08 - INFO - codeparrot_training - Step 15072: {'lr': 0.0004139488864212578, 'samples': 2894016, 'steps': 15072, 'loss/train': 0.9070402085781097} +01/27/2022 10:02:11 - INFO - codeparrot_training - Step 15073: {'lr': 0.0004139365334171059, 'samples': 2894208, 'steps': 15073, 'loss/train': 0.7397717088460922} +01/27/2022 10:02:15 - INFO - codeparrot_training - Step 15074: {'lr': 0.0004139241797107019, 'samples': 2894400, 'steps': 15074, 'loss/train': 0.69002665579319} +01/27/2022 10:02:18 - INFO - codeparrot_training - Step 15075: {'lr': 0.00041391182530209873, 'samples': 2894592, 'steps': 15075, 'loss/train': 0.7817244529724121} +01/27/2022 10:02:21 - INFO - codeparrot_training - Step 15076: {'lr': 0.0004138994701913494, 'samples': 2894784, 'steps': 15076, 'loss/train': 0.8523864448070526} +01/27/2022 10:02:24 - INFO - codeparrot_training - Step 15077: {'lr': 0.00041388711437850676, 'samples': 2894976, 'steps': 15077, 'loss/train': 0.7179430425167084} +01/27/2022 10:02:27 - INFO - codeparrot_training - Step 15078: {'lr': 0.00041387475786362386, 'samples': 2895168, 'steps': 15078, 'loss/train': 0.7193187922239304} +01/27/2022 10:02:30 - INFO - codeparrot_training - Step 15079: {'lr': 0.0004138624006467534, 'samples': 2895360, 'steps': 15079, 'loss/train': 0.8820582032203674} +01/27/2022 10:02:34 - INFO - codeparrot_training - Step 15080: {'lr': 0.00041385004272794846, 'samples': 2895552, 'steps': 15080, 'loss/train': 0.695886567234993} +01/27/2022 10:02:40 - INFO - codeparrot_training - Step 15081: {'lr': 0.00041383768410726207, 'samples': 2895744, 'steps': 15081, 'loss/train': 1.1517982184886932} +01/27/2022 10:02:43 - INFO - codeparrot_training - Step 15082: {'lr': 0.000413825324784747, 'samples': 2895936, 'steps': 15082, 'loss/train': 1.1338875889778137} +01/27/2022 10:02:46 - INFO - codeparrot_training - Step 15083: {'lr': 0.00041381296476045626, 'samples': 2896128, 'steps': 15083, 'loss/train': 0.8780824542045593} +01/27/2022 10:02:49 - INFO - codeparrot_training - Step 15084: {'lr': 0.0004138006040344428, 'samples': 2896320, 'steps': 15084, 'loss/train': 0.6773297935724258} +01/27/2022 10:02:52 - INFO - codeparrot_training - Step 15085: {'lr': 0.0004137882426067595, 'samples': 2896512, 'steps': 15085, 'loss/train': 0.568618431687355} +01/27/2022 10:02:55 - INFO - codeparrot_training - Step 15086: {'lr': 0.0004137758804774594, 'samples': 2896704, 'steps': 15086, 'loss/train': 1.054057538509369} +01/27/2022 10:02:59 - INFO - codeparrot_training - Step 15087: {'lr': 0.0004137635176465955, 'samples': 2896896, 'steps': 15087, 'loss/train': 1.364462435245514} +01/27/2022 10:03:02 - INFO - codeparrot_training - Step 15088: {'lr': 0.00041375115411422064, 'samples': 2897088, 'steps': 15088, 'loss/train': 1.2102851271629333} +01/27/2022 10:03:07 - INFO - codeparrot_training - Step 15089: {'lr': 0.0004137387898803878, 'samples': 2897280, 'steps': 15089, 'loss/train': 0.9780047535896301} +01/27/2022 10:03:10 - INFO - codeparrot_training - Step 15090: {'lr': 0.0004137264249451501, 'samples': 2897472, 'steps': 15090, 'loss/train': 0.7758292257785797} +01/27/2022 10:03:13 - INFO - codeparrot_training - Step 15091: {'lr': 0.00041371405930856026, 'samples': 2897664, 'steps': 15091, 'loss/train': 0.9431616961956024} +01/27/2022 10:03:16 - INFO - codeparrot_training - Step 15092: {'lr': 0.00041370169297067145, 'samples': 2897856, 'steps': 15092, 'loss/train': 1.7366500496864319} +01/27/2022 10:03:19 - INFO - codeparrot_training - Step 15093: {'lr': 0.0004136893259315365, 'samples': 2898048, 'steps': 15093, 'loss/train': 0.8432925045490265} +01/27/2022 10:03:22 - INFO - codeparrot_training - Step 15094: {'lr': 0.00041367695819120854, 'samples': 2898240, 'steps': 15094, 'loss/train': 0.6958901435136795} +01/27/2022 10:03:26 - INFO - codeparrot_training - Step 15095: {'lr': 0.0004136645897497404, 'samples': 2898432, 'steps': 15095, 'loss/train': 0.5880395472049713} +01/27/2022 10:03:29 - INFO - codeparrot_training - Step 15096: {'lr': 0.0004136522206071852, 'samples': 2898624, 'steps': 15096, 'loss/train': 1.138904482126236} +01/27/2022 10:03:32 - INFO - codeparrot_training - Step 15097: {'lr': 0.0004136398507635958, 'samples': 2898816, 'steps': 15097, 'loss/train': 0.9278659522533417} +01/27/2022 10:03:35 - INFO - codeparrot_training - Step 15098: {'lr': 0.00041362748021902526, 'samples': 2899008, 'steps': 15098, 'loss/train': 0.12431562691926956} +01/27/2022 10:03:41 - INFO - codeparrot_training - Step 15099: {'lr': 0.0004136151089735265, 'samples': 2899200, 'steps': 15099, 'loss/train': 1.3303684294223785} +01/27/2022 10:03:44 - INFO - codeparrot_training - Step 15100: {'lr': 0.00041360273702715263, 'samples': 2899392, 'steps': 15100, 'loss/train': 0.8454715311527252} +01/27/2022 10:03:48 - INFO - codeparrot_training - Step 15101: {'lr': 0.0004135903643799566, 'samples': 2899584, 'steps': 15101, 'loss/train': 1.3062570691108704} +01/27/2022 10:03:51 - INFO - codeparrot_training - Step 15102: {'lr': 0.00041357799103199127, 'samples': 2899776, 'steps': 15102, 'loss/train': 0.3033187910914421} +01/27/2022 10:03:54 - INFO - codeparrot_training - Step 15103: {'lr': 0.00041356561698330984, 'samples': 2899968, 'steps': 15103, 'loss/train': 0.6348081529140472} +01/27/2022 10:03:57 - INFO - codeparrot_training - Step 15104: {'lr': 0.0004135532422339653, 'samples': 2900160, 'steps': 15104, 'loss/train': 0.9333213865756989} +01/27/2022 10:04:00 - INFO - codeparrot_training - Step 15105: {'lr': 0.00041354086678401056, 'samples': 2900352, 'steps': 15105, 'loss/train': 1.0410080552101135} +01/27/2022 10:04:03 - INFO - codeparrot_training - Step 15106: {'lr': 0.00041352849063349865, 'samples': 2900544, 'steps': 15106, 'loss/train': 0.6982188820838928} +01/27/2022 10:04:06 - INFO - codeparrot_training - Step 15107: {'lr': 0.0004135161137824827, 'samples': 2900736, 'steps': 15107, 'loss/train': 0.6585945636034012} +01/27/2022 10:04:11 - INFO - codeparrot_training - Step 15108: {'lr': 0.0004135037362310155, 'samples': 2900928, 'steps': 15108, 'loss/train': 0.29781322181224823} +01/27/2022 10:04:14 - INFO - codeparrot_training - Step 15109: {'lr': 0.0004134913579791503, 'samples': 2901120, 'steps': 15109, 'loss/train': 0.4704756885766983} +01/27/2022 10:04:17 - INFO - codeparrot_training - Step 15110: {'lr': 0.00041347897902694003, 'samples': 2901312, 'steps': 15110, 'loss/train': 0.14110201969742775} +01/27/2022 10:04:20 - INFO - codeparrot_training - Step 15111: {'lr': 0.00041346659937443775, 'samples': 2901504, 'steps': 15111, 'loss/train': 0.6410219371318817} +01/27/2022 10:04:23 - INFO - codeparrot_training - Step 15112: {'lr': 0.00041345421902169645, 'samples': 2901696, 'steps': 15112, 'loss/train': 0.9590623676776886} +01/27/2022 10:04:26 - INFO - codeparrot_training - Step 15113: {'lr': 0.0004134418379687691, 'samples': 2901888, 'steps': 15113, 'loss/train': 0.3362956792116165} +01/27/2022 10:04:30 - INFO - codeparrot_training - Step 15114: {'lr': 0.0004134294562157089, 'samples': 2902080, 'steps': 15114, 'loss/train': 0.8609730005264282} +01/27/2022 10:04:33 - INFO - codeparrot_training - Step 15115: {'lr': 0.00041341707376256877, 'samples': 2902272, 'steps': 15115, 'loss/train': 1.0970311760902405} +01/27/2022 10:04:37 - INFO - codeparrot_training - Step 15116: {'lr': 0.00041340469060940183, 'samples': 2902464, 'steps': 15116, 'loss/train': 0.9373852014541626} +01/27/2022 10:04:40 - INFO - codeparrot_training - Step 15117: {'lr': 0.0004133923067562611, 'samples': 2902656, 'steps': 15117, 'loss/train': 0.29573075473308563} +01/27/2022 10:04:44 - INFO - codeparrot_training - Step 15118: {'lr': 0.0004133799222031995, 'samples': 2902848, 'steps': 15118, 'loss/train': 0.5302037000656128} +01/27/2022 10:04:47 - INFO - codeparrot_training - Step 15119: {'lr': 0.0004133675369502703, 'samples': 2903040, 'steps': 15119, 'loss/train': 0.5095786303281784} +01/27/2022 10:04:50 - INFO - codeparrot_training - Step 15120: {'lr': 0.0004133551509975264, 'samples': 2903232, 'steps': 15120, 'loss/train': 0.9138329029083252} +01/27/2022 10:04:53 - INFO - codeparrot_training - Step 15121: {'lr': 0.0004133427643450209, 'samples': 2903424, 'steps': 15121, 'loss/train': 1.2112557291984558} +01/27/2022 10:04:56 - INFO - codeparrot_training - Step 15122: {'lr': 0.0004133303769928068, 'samples': 2903616, 'steps': 15122, 'loss/train': 1.0615377724170685} +01/27/2022 10:04:59 - INFO - codeparrot_training - Step 15123: {'lr': 0.00041331798894093735, 'samples': 2903808, 'steps': 15123, 'loss/train': 0.45769791305065155} +01/27/2022 10:05:02 - INFO - codeparrot_training - Step 15124: {'lr': 0.0004133056001894655, 'samples': 2904000, 'steps': 15124, 'loss/train': 0.7500764429569244} +01/27/2022 10:05:07 - INFO - codeparrot_training - Step 15125: {'lr': 0.0004132932107384442, 'samples': 2904192, 'steps': 15125, 'loss/train': 0.7160933464765549} +01/27/2022 10:05:10 - INFO - codeparrot_training - Step 15126: {'lr': 0.0004132808205879267, 'samples': 2904384, 'steps': 15126, 'loss/train': 0.1793643683195114} +01/27/2022 10:05:13 - INFO - codeparrot_training - Step 15127: {'lr': 0.000413268429737966, 'samples': 2904576, 'steps': 15127, 'loss/train': 0.4053032398223877} +01/27/2022 10:05:16 - INFO - codeparrot_training - Step 15128: {'lr': 0.00041325603818861517, 'samples': 2904768, 'steps': 15128, 'loss/train': 0.7500234246253967} +01/27/2022 10:05:20 - INFO - codeparrot_training - Step 15129: {'lr': 0.00041324364593992735, 'samples': 2904960, 'steps': 15129, 'loss/train': 1.0157876908779144} +01/27/2022 10:05:23 - INFO - codeparrot_training - Step 15130: {'lr': 0.00041323125299195563, 'samples': 2905152, 'steps': 15130, 'loss/train': 0.9396091997623444} +01/27/2022 10:05:26 - INFO - codeparrot_training - Step 15131: {'lr': 0.000413218859344753, 'samples': 2905344, 'steps': 15131, 'loss/train': 0.9353137314319611} +01/27/2022 10:05:29 - INFO - codeparrot_training - Step 15132: {'lr': 0.00041320646499837254, 'samples': 2905536, 'steps': 15132, 'loss/train': 1.380929410457611} +01/27/2022 10:05:32 - INFO - codeparrot_training - Step 15133: {'lr': 0.00041319406995286753, 'samples': 2905728, 'steps': 15133, 'loss/train': 0.6920620501041412} +01/27/2022 10:05:38 - INFO - codeparrot_training - Step 15134: {'lr': 0.0004131816742082909, 'samples': 2905920, 'steps': 15134, 'loss/train': 1.0013879835605621} +01/27/2022 10:05:42 - INFO - codeparrot_training - Step 15135: {'lr': 0.00041316927776469575, 'samples': 2906112, 'steps': 15135, 'loss/train': 0.948353111743927} +01/27/2022 10:05:45 - INFO - codeparrot_training - Step 15136: {'lr': 0.00041315688062213524, 'samples': 2906304, 'steps': 15136, 'loss/train': 0.7306952029466629} +01/27/2022 10:05:48 - INFO - codeparrot_training - Step 15137: {'lr': 0.0004131444827806625, 'samples': 2906496, 'steps': 15137, 'loss/train': 0.7089412361383438} +01/27/2022 10:05:51 - INFO - codeparrot_training - Step 15138: {'lr': 0.00041313208424033056, 'samples': 2906688, 'steps': 15138, 'loss/train': 0.522646889090538} +01/27/2022 10:05:54 - INFO - codeparrot_training - Step 15139: {'lr': 0.0004131196850011926, 'samples': 2906880, 'steps': 15139, 'loss/train': 0.9984056353569031} +01/27/2022 10:05:57 - INFO - codeparrot_training - Step 15140: {'lr': 0.0004131072850633017, 'samples': 2907072, 'steps': 15140, 'loss/train': 0.9814590811729431} +01/27/2022 10:06:00 - INFO - codeparrot_training - Step 15141: {'lr': 0.00041309488442671093, 'samples': 2907264, 'steps': 15141, 'loss/train': 0.6543077677488327} +01/27/2022 10:06:04 - INFO - codeparrot_training - Step 15142: {'lr': 0.00041308248309147356, 'samples': 2907456, 'steps': 15142, 'loss/train': 0.7398658096790314} +01/27/2022 10:06:08 - INFO - codeparrot_training - Step 15143: {'lr': 0.00041307008105764256, 'samples': 2907648, 'steps': 15143, 'loss/train': 0.8784997165203094} +01/27/2022 10:06:11 - INFO - codeparrot_training - Step 15144: {'lr': 0.0004130576783252712, 'samples': 2907840, 'steps': 15144, 'loss/train': 0.7648639976978302} +01/27/2022 10:06:14 - INFO - codeparrot_training - Step 15145: {'lr': 0.00041304527489441237, 'samples': 2908032, 'steps': 15145, 'loss/train': 0.8827351927757263} +01/27/2022 10:06:18 - INFO - codeparrot_training - Step 15146: {'lr': 0.0004130328707651195, 'samples': 2908224, 'steps': 15146, 'loss/train': 0.8313217163085938} +01/27/2022 10:06:21 - INFO - codeparrot_training - Step 15147: {'lr': 0.00041302046593744547, 'samples': 2908416, 'steps': 15147, 'loss/train': 1.0079635977745056} +01/27/2022 10:06:24 - INFO - codeparrot_training - Step 15148: {'lr': 0.00041300806041144356, 'samples': 2908608, 'steps': 15148, 'loss/train': 0.8122485280036926} +01/27/2022 10:06:27 - INFO - codeparrot_training - Step 15149: {'lr': 0.0004129956541871669, 'samples': 2908800, 'steps': 15149, 'loss/train': 0.8511426150798798} +01/27/2022 10:06:30 - INFO - codeparrot_training - Step 15150: {'lr': 0.00041298324726466855, 'samples': 2908992, 'steps': 15150, 'loss/train': 1.2348343133926392} +01/27/2022 10:06:33 - INFO - codeparrot_training - Step 15151: {'lr': 0.0004129708396440018, 'samples': 2909184, 'steps': 15151, 'loss/train': 0.4877796918153763} +01/27/2022 10:06:38 - INFO - codeparrot_training - Step 15152: {'lr': 0.00041295843132521973, 'samples': 2909376, 'steps': 15152, 'loss/train': 0.6658680886030197} +01/27/2022 10:06:41 - INFO - codeparrot_training - Step 15153: {'lr': 0.0004129460223083754, 'samples': 2909568, 'steps': 15153, 'loss/train': 0.5225119292736053} +01/27/2022 10:06:44 - INFO - codeparrot_training - Step 15154: {'lr': 0.0004129336125935221, 'samples': 2909760, 'steps': 15154, 'loss/train': 0.78013014793396} +01/27/2022 10:06:47 - INFO - codeparrot_training - Step 15155: {'lr': 0.000412921202180713, 'samples': 2909952, 'steps': 15155, 'loss/train': 0.5048753321170807} +01/27/2022 10:06:50 - INFO - codeparrot_training - Step 15156: {'lr': 0.00041290879107000114, 'samples': 2910144, 'steps': 15156, 'loss/train': 0.9795485436916351} +01/27/2022 10:06:53 - INFO - codeparrot_training - Step 15157: {'lr': 0.00041289637926143974, 'samples': 2910336, 'steps': 15157, 'loss/train': 0.7698147296905518} +01/27/2022 10:06:57 - INFO - codeparrot_training - Step 15158: {'lr': 0.000412883966755082, 'samples': 2910528, 'steps': 15158, 'loss/train': 0.6363487243652344} +01/27/2022 10:07:00 - INFO - codeparrot_training - Step 15159: {'lr': 0.000412871553550981, 'samples': 2910720, 'steps': 15159, 'loss/train': 1.1220355331897736} +01/27/2022 10:07:03 - INFO - codeparrot_training - Step 15160: {'lr': 0.00041285913964919006, 'samples': 2910912, 'steps': 15160, 'loss/train': 0.5562449544668198} +01/27/2022 10:07:09 - INFO - codeparrot_training - Step 15161: {'lr': 0.0004128467250497623, 'samples': 2911104, 'steps': 15161, 'loss/train': 1.2295538485050201} +01/27/2022 10:07:12 - INFO - codeparrot_training - Step 15162: {'lr': 0.00041283430975275085, 'samples': 2911296, 'steps': 15162, 'loss/train': 0.7895789444446564} +01/27/2022 10:07:15 - INFO - codeparrot_training - Step 15163: {'lr': 0.0004128218937582089, 'samples': 2911488, 'steps': 15163, 'loss/train': 0.8025662004947662} +01/27/2022 10:07:18 - INFO - codeparrot_training - Step 15164: {'lr': 0.00041280947706618965, 'samples': 2911680, 'steps': 15164, 'loss/train': 0.8695378303527832} +01/27/2022 10:07:22 - INFO - codeparrot_training - Step 15165: {'lr': 0.00041279705967674636, 'samples': 2911872, 'steps': 15165, 'loss/train': 0.7063073068857193} +01/27/2022 10:07:25 - INFO - codeparrot_training - Step 15166: {'lr': 0.00041278464158993214, 'samples': 2912064, 'steps': 15166, 'loss/train': 0.5916159152984619} +01/27/2022 10:07:28 - INFO - codeparrot_training - Step 15167: {'lr': 0.0004127722228058002, 'samples': 2912256, 'steps': 15167, 'loss/train': 0.1484822742640972} +01/27/2022 10:07:31 - INFO - codeparrot_training - Step 15168: {'lr': 0.0004127598033244037, 'samples': 2912448, 'steps': 15168, 'loss/train': 0.8348661661148071} +01/27/2022 10:07:36 - INFO - codeparrot_training - Step 15169: {'lr': 0.0004127473831457959, 'samples': 2912640, 'steps': 15169, 'loss/train': 0.6404544711112976} +01/27/2022 10:07:39 - INFO - codeparrot_training - Step 15170: {'lr': 0.00041273496227003004, 'samples': 2912832, 'steps': 15170, 'loss/train': 0.42868445813655853} +01/27/2022 10:07:43 - INFO - codeparrot_training - Step 15171: {'lr': 0.0004127225406971592, 'samples': 2913024, 'steps': 15171, 'loss/train': 0.03490063827484846} +01/27/2022 10:07:46 - INFO - codeparrot_training - Step 15172: {'lr': 0.00041271011842723676, 'samples': 2913216, 'steps': 15172, 'loss/train': 0.6420041620731354} +01/27/2022 10:07:49 - INFO - codeparrot_training - Step 15173: {'lr': 0.00041269769546031576, 'samples': 2913408, 'steps': 15173, 'loss/train': 0.8425111770629883} +01/27/2022 10:07:52 - INFO - codeparrot_training - Step 15174: {'lr': 0.0004126852717964495, 'samples': 2913600, 'steps': 15174, 'loss/train': 1.9573153853416443} +01/27/2022 10:07:55 - INFO - codeparrot_training - Step 15175: {'lr': 0.0004126728474356912, 'samples': 2913792, 'steps': 15175, 'loss/train': 0.7955903112888336} +01/27/2022 10:07:58 - INFO - codeparrot_training - Step 15176: {'lr': 0.0004126604223780941, 'samples': 2913984, 'steps': 15176, 'loss/train': 1.7590491771697998} +01/27/2022 10:08:01 - INFO - codeparrot_training - Step 15177: {'lr': 0.00041264799662371144, 'samples': 2914176, 'steps': 15177, 'loss/train': 0.7702573835849762} +01/27/2022 10:08:05 - INFO - codeparrot_training - Step 15178: {'lr': 0.0004126355701725963, 'samples': 2914368, 'steps': 15178, 'loss/train': 1.3199823796749115} +01/27/2022 10:08:11 - INFO - codeparrot_training - Step 15179: {'lr': 0.00041262314302480216, 'samples': 2914560, 'steps': 15179, 'loss/train': 0.39415620267391205} +01/27/2022 10:08:14 - INFO - codeparrot_training - Step 15180: {'lr': 0.000412610715180382, 'samples': 2914752, 'steps': 15180, 'loss/train': 0.6454146355390549} +01/27/2022 10:08:17 - INFO - codeparrot_training - Step 15181: {'lr': 0.0004125982866393892, 'samples': 2914944, 'steps': 15181, 'loss/train': 0.9572467803955078} +01/27/2022 10:08:20 - INFO - codeparrot_training - Step 15182: {'lr': 0.0004125858574018769, 'samples': 2915136, 'steps': 15182, 'loss/train': 0.5039267241954803} +01/27/2022 10:08:23 - INFO - codeparrot_training - Step 15183: {'lr': 0.0004125734274678986, 'samples': 2915328, 'steps': 15183, 'loss/train': 0.20860301703214645} +01/27/2022 10:08:26 - INFO - codeparrot_training - Step 15184: {'lr': 0.0004125609968375072, 'samples': 2915520, 'steps': 15184, 'loss/train': 1.0032720565795898} +01/27/2022 10:08:30 - INFO - codeparrot_training - Step 15185: {'lr': 0.00041254856551075616, 'samples': 2915712, 'steps': 15185, 'loss/train': 0.7500565946102142} +01/27/2022 10:08:33 - INFO - codeparrot_training - Step 15186: {'lr': 0.0004125361334876987, 'samples': 2915904, 'steps': 15186, 'loss/train': 0.8125545680522919} +01/27/2022 10:08:36 - INFO - codeparrot_training - Step 15187: {'lr': 0.000412523700768388, 'samples': 2916096, 'steps': 15187, 'loss/train': 0.8894788026809692} +01/27/2022 10:08:40 - INFO - codeparrot_training - Step 15188: {'lr': 0.0004125112673528775, 'samples': 2916288, 'steps': 15188, 'loss/train': 1.149211585521698} +01/27/2022 10:08:43 - INFO - codeparrot_training - Step 15189: {'lr': 0.0004124988332412202, 'samples': 2916480, 'steps': 15189, 'loss/train': 1.0949885845184326} +01/27/2022 10:08:46 - INFO - codeparrot_training - Step 15190: {'lr': 0.00041248639843346953, 'samples': 2916672, 'steps': 15190, 'loss/train': 0.6252223402261734} +01/27/2022 10:08:50 - INFO - codeparrot_training - Step 15191: {'lr': 0.0004124739629296787, 'samples': 2916864, 'steps': 15191, 'loss/train': 1.060896635055542} +01/27/2022 10:08:53 - INFO - codeparrot_training - Step 15192: {'lr': 0.00041246152672990105, 'samples': 2917056, 'steps': 15192, 'loss/train': 0.7346066236495972} +01/27/2022 10:08:56 - INFO - codeparrot_training - Step 15193: {'lr': 0.00041244908983418985, 'samples': 2917248, 'steps': 15193, 'loss/train': 0.8659083545207977} +01/27/2022 10:08:59 - INFO - codeparrot_training - Step 15194: {'lr': 0.0004124366522425982, 'samples': 2917440, 'steps': 15194, 'loss/train': 1.2697033882141113} +01/27/2022 10:09:02 - INFO - codeparrot_training - Step 15195: {'lr': 0.0004124242139551796, 'samples': 2917632, 'steps': 15195, 'loss/train': 1.1108210384845734} +01/27/2022 10:09:05 - INFO - codeparrot_training - Step 15196: {'lr': 0.00041241177497198725, 'samples': 2917824, 'steps': 15196, 'loss/train': 0.9862688183784485} +01/27/2022 10:09:11 - INFO - codeparrot_training - Step 15197: {'lr': 0.00041239933529307437, 'samples': 2918016, 'steps': 15197, 'loss/train': 0.6750245690345764} +01/27/2022 10:09:14 - INFO - codeparrot_training - Step 15198: {'lr': 0.00041238689491849434, 'samples': 2918208, 'steps': 15198, 'loss/train': 1.0515141785144806} +01/27/2022 10:09:18 - INFO - codeparrot_training - Step 15199: {'lr': 0.00041237445384830043, 'samples': 2918400, 'steps': 15199, 'loss/train': 1.4679451882839203} +01/27/2022 10:09:21 - INFO - codeparrot_training - Step 15200: {'lr': 0.0004123620120825459, 'samples': 2918592, 'steps': 15200, 'loss/train': 1.0032333433628082} +01/27/2022 10:09:24 - INFO - codeparrot_training - Step 15201: {'lr': 0.0004123495696212841, 'samples': 2918784, 'steps': 15201, 'loss/train': 0.8444022238254547} +01/27/2022 10:09:27 - INFO - codeparrot_training - Step 15202: {'lr': 0.00041233712646456823, 'samples': 2918976, 'steps': 15202, 'loss/train': 0.5353946685791016} +01/27/2022 10:09:30 - INFO - codeparrot_training - Step 15203: {'lr': 0.0004123246826124517, 'samples': 2919168, 'steps': 15203, 'loss/train': 0.5479595214128494} +01/27/2022 10:09:33 - INFO - codeparrot_training - Step 15204: {'lr': 0.00041231223806498777, 'samples': 2919360, 'steps': 15204, 'loss/train': 0.7386463433504105} +01/27/2022 10:09:38 - INFO - codeparrot_training - Step 15205: {'lr': 0.0004122997928222298, 'samples': 2919552, 'steps': 15205, 'loss/train': 1.018709510564804} +01/27/2022 10:09:41 - INFO - codeparrot_training - Step 15206: {'lr': 0.000412287346884231, 'samples': 2919744, 'steps': 15206, 'loss/train': 1.233458787202835} +01/27/2022 10:09:44 - INFO - codeparrot_training - Step 15207: {'lr': 0.00041227490025104474, 'samples': 2919936, 'steps': 15207, 'loss/train': 0.904532790184021} +01/27/2022 10:09:47 - INFO - codeparrot_training - Step 15208: {'lr': 0.00041226245292272433, 'samples': 2920128, 'steps': 15208, 'loss/train': 0.3402152583003044} +01/27/2022 10:09:50 - INFO - codeparrot_training - Step 15209: {'lr': 0.00041225000489932315, 'samples': 2920320, 'steps': 15209, 'loss/train': 0.8034434616565704} +01/27/2022 10:09:53 - INFO - codeparrot_training - Step 15210: {'lr': 0.00041223755618089445, 'samples': 2920512, 'steps': 15210, 'loss/train': 1.1959684789180756} +01/27/2022 10:09:56 - INFO - codeparrot_training - Step 15211: {'lr': 0.0004122251067674915, 'samples': 2920704, 'steps': 15211, 'loss/train': 0.8464322984218597} +01/27/2022 10:10:00 - INFO - codeparrot_training - Step 15212: {'lr': 0.00041221265665916776, 'samples': 2920896, 'steps': 15212, 'loss/train': 0.2642107307910919} +01/27/2022 10:10:03 - INFO - codeparrot_training - Step 15213: {'lr': 0.0004122002058559765, 'samples': 2921088, 'steps': 15213, 'loss/train': 0.8393420577049255} +01/27/2022 10:10:07 - INFO - codeparrot_training - Step 15214: {'lr': 0.00041218775435797106, 'samples': 2921280, 'steps': 15214, 'loss/train': 1.1217803657054901} +01/27/2022 10:10:11 - INFO - codeparrot_training - Step 15215: {'lr': 0.0004121753021652048, 'samples': 2921472, 'steps': 15215, 'loss/train': 0.7977834641933441} +01/27/2022 10:10:14 - INFO - codeparrot_training - Step 15216: {'lr': 0.0004121628492777311, 'samples': 2921664, 'steps': 15216, 'loss/train': 0.8108708560466766} +01/27/2022 10:10:17 - INFO - codeparrot_training - Step 15217: {'lr': 0.0004121503956956031, 'samples': 2921856, 'steps': 15217, 'loss/train': 0.8680626153945923} +01/27/2022 10:10:20 - INFO - codeparrot_training - Step 15218: {'lr': 0.0004121379414188744, 'samples': 2922048, 'steps': 15218, 'loss/train': 0.6437699943780899} +01/27/2022 10:10:23 - INFO - codeparrot_training - Step 15219: {'lr': 0.0004121254864475982, 'samples': 2922240, 'steps': 15219, 'loss/train': 0.0846392959356308} +01/27/2022 10:10:26 - INFO - codeparrot_training - Step 15220: {'lr': 0.0004121130307818279, 'samples': 2922432, 'steps': 15220, 'loss/train': 0.8900691568851471} +01/27/2022 10:10:29 - INFO - codeparrot_training - Step 15221: {'lr': 0.00041210057442161687, 'samples': 2922624, 'steps': 15221, 'loss/train': 1.0904810428619385} +01/27/2022 10:10:33 - INFO - codeparrot_training - Step 15222: {'lr': 0.0004120881173670184, 'samples': 2922816, 'steps': 15222, 'loss/train': 0.8718493580818176} +01/27/2022 10:10:38 - INFO - codeparrot_training - Step 15223: {'lr': 0.000412075659618086, 'samples': 2923008, 'steps': 15223, 'loss/train': 0.9674663543701172} +01/27/2022 10:10:41 - INFO - codeparrot_training - Step 15224: {'lr': 0.0004120632011748728, 'samples': 2923200, 'steps': 15224, 'loss/train': 0.3632773831486702} +01/27/2022 10:10:44 - INFO - codeparrot_training - Step 15225: {'lr': 0.00041205074203743244, 'samples': 2923392, 'steps': 15225, 'loss/train': 0.8264859616756439} +01/27/2022 10:10:47 - INFO - codeparrot_training - Step 15226: {'lr': 0.00041203828220581805, 'samples': 2923584, 'steps': 15226, 'loss/train': 0.7806507647037506} +01/27/2022 10:10:50 - INFO - codeparrot_training - Step 15227: {'lr': 0.00041202582168008324, 'samples': 2923776, 'steps': 15227, 'loss/train': 0.6771206259727478} +01/27/2022 10:10:53 - INFO - codeparrot_training - Step 15228: {'lr': 0.00041201336046028117, 'samples': 2923968, 'steps': 15228, 'loss/train': 0.7714901268482208} +01/27/2022 10:10:56 - INFO - codeparrot_training - Step 15229: {'lr': 0.0004120008985464654, 'samples': 2924160, 'steps': 15229, 'loss/train': 0.831468254327774} +01/27/2022 10:11:00 - INFO - codeparrot_training - Step 15230: {'lr': 0.0004119884359386891, 'samples': 2924352, 'steps': 15230, 'loss/train': 0.4272100031375885} +01/27/2022 10:11:03 - INFO - codeparrot_training - Step 15231: {'lr': 0.0004119759726370058, 'samples': 2924544, 'steps': 15231, 'loss/train': 0.5351268500089645} +01/27/2022 10:11:07 - INFO - codeparrot_training - Step 15232: {'lr': 0.0004119635086414689, 'samples': 2924736, 'steps': 15232, 'loss/train': 0.843328446149826} +01/27/2022 10:11:11 - INFO - codeparrot_training - Step 15233: {'lr': 0.0004119510439521318, 'samples': 2924928, 'steps': 15233, 'loss/train': 0.7913035154342651} +01/27/2022 10:11:14 - INFO - codeparrot_training - Step 15234: {'lr': 0.0004119385785690478, 'samples': 2925120, 'steps': 15234, 'loss/train': 0.8355817794799805} +01/27/2022 10:11:17 - INFO - codeparrot_training - Step 15235: {'lr': 0.0004119261124922703, 'samples': 2925312, 'steps': 15235, 'loss/train': 1.0629558563232422} +01/27/2022 10:11:20 - INFO - codeparrot_training - Step 15236: {'lr': 0.00041191364572185286, 'samples': 2925504, 'steps': 15236, 'loss/train': 0.8704325258731842} +01/27/2022 10:11:23 - INFO - codeparrot_training - Step 15237: {'lr': 0.0004119011782578487, 'samples': 2925696, 'steps': 15237, 'loss/train': 0.3841683268547058} +01/27/2022 10:11:26 - INFO - codeparrot_training - Step 15238: {'lr': 0.00041188871010031135, 'samples': 2925888, 'steps': 15238, 'loss/train': 0.9356585741043091} +01/27/2022 10:11:30 - INFO - codeparrot_training - Step 15239: {'lr': 0.0004118762412492941, 'samples': 2926080, 'steps': 15239, 'loss/train': 0.7363952547311783} +01/27/2022 10:11:33 - INFO - codeparrot_training - Step 15240: {'lr': 0.00041186377170485057, 'samples': 2926272, 'steps': 15240, 'loss/train': 0.823079913854599} +01/27/2022 10:11:37 - INFO - codeparrot_training - Step 15241: {'lr': 0.00041185130146703387, 'samples': 2926464, 'steps': 15241, 'loss/train': 0.94053053855896} +01/27/2022 10:11:40 - INFO - codeparrot_training - Step 15242: {'lr': 0.0004118388305358977, 'samples': 2926656, 'steps': 15242, 'loss/train': 0.7145417332649231} +01/27/2022 10:11:43 - INFO - codeparrot_training - Step 15243: {'lr': 0.0004118263589114953, 'samples': 2926848, 'steps': 15243, 'loss/train': 0.7779580950737} +01/27/2022 10:11:47 - INFO - codeparrot_training - Step 15244: {'lr': 0.00041181388659388026, 'samples': 2927040, 'steps': 15244, 'loss/train': 1.0018833875656128} +01/27/2022 10:11:50 - INFO - codeparrot_training - Step 15245: {'lr': 0.00041180141358310586, 'samples': 2927232, 'steps': 15245, 'loss/train': 0.9527736604213715} +01/27/2022 10:11:53 - INFO - codeparrot_training - Step 15246: {'lr': 0.00041178893987922556, 'samples': 2927424, 'steps': 15246, 'loss/train': 0.579461395740509} +01/27/2022 10:11:56 - INFO - codeparrot_training - Step 15247: {'lr': 0.0004117764654822929, 'samples': 2927616, 'steps': 15247, 'loss/train': 0.7173086553812027} +01/27/2022 10:11:59 - INFO - codeparrot_training - Step 15248: {'lr': 0.0004117639903923611, 'samples': 2927808, 'steps': 15248, 'loss/train': 0.7694372534751892} +01/27/2022 10:12:04 - INFO - codeparrot_training - Step 15249: {'lr': 0.0004117515146094838, 'samples': 2928000, 'steps': 15249, 'loss/train': 0.533294141292572} +01/27/2022 10:12:08 - INFO - codeparrot_training - Step 15250: {'lr': 0.0004117390381337144, 'samples': 2928192, 'steps': 15250, 'loss/train': 0.7289541810750961} +01/27/2022 10:12:11 - INFO - codeparrot_training - Step 15251: {'lr': 0.00041172656096510624, 'samples': 2928384, 'steps': 15251, 'loss/train': 0.7778588533401489} +01/27/2022 10:12:14 - INFO - codeparrot_training - Step 15252: {'lr': 0.0004117140831037129, 'samples': 2928576, 'steps': 15252, 'loss/train': 0.36101067066192627} +01/27/2022 10:12:17 - INFO - codeparrot_training - Step 15253: {'lr': 0.00041170160454958785, 'samples': 2928768, 'steps': 15253, 'loss/train': 0.9255006909370422} +01/27/2022 10:12:20 - INFO - codeparrot_training - Step 15254: {'lr': 0.00041168912530278434, 'samples': 2928960, 'steps': 15254, 'loss/train': 0.7794992029666901} +01/27/2022 10:12:23 - INFO - codeparrot_training - Step 15255: {'lr': 0.00041167664536335605, 'samples': 2929152, 'steps': 15255, 'loss/train': 0.8711095154285431} +01/27/2022 10:12:26 - INFO - codeparrot_training - Step 15256: {'lr': 0.0004116641647313563, 'samples': 2929344, 'steps': 15256, 'loss/train': 0.9579510390758514} +01/27/2022 10:12:30 - INFO - codeparrot_training - Step 15257: {'lr': 0.00041165168340683857, 'samples': 2929536, 'steps': 15257, 'loss/train': 1.2563281059265137} +01/27/2022 10:12:34 - INFO - codeparrot_training - Step 15258: {'lr': 0.0004116392013898564, 'samples': 2929728, 'steps': 15258, 'loss/train': 1.565624177455902} +01/27/2022 10:12:37 - INFO - codeparrot_training - Step 15259: {'lr': 0.0004116267186804632, 'samples': 2929920, 'steps': 15259, 'loss/train': 1.0207874178886414} +01/27/2022 10:12:40 - INFO - codeparrot_training - Step 15260: {'lr': 0.0004116142352787125, 'samples': 2930112, 'steps': 15260, 'loss/train': 0.8236487209796906} +01/27/2022 10:12:43 - INFO - codeparrot_training - Step 15261: {'lr': 0.0004116017511846577, 'samples': 2930304, 'steps': 15261, 'loss/train': 0.8386998474597931} +01/27/2022 10:12:47 - INFO - codeparrot_training - Step 15262: {'lr': 0.00041158926639835234, 'samples': 2930496, 'steps': 15262, 'loss/train': 0.7969809472560883} +01/27/2022 10:12:50 - INFO - codeparrot_training - Step 15263: {'lr': 0.00041157678091984987, 'samples': 2930688, 'steps': 15263, 'loss/train': 0.9494924247264862} +01/27/2022 10:12:53 - INFO - codeparrot_training - Step 15264: {'lr': 0.0004115642947492038, 'samples': 2930880, 'steps': 15264, 'loss/train': 1.0654816925525665} +01/27/2022 10:12:56 - INFO - codeparrot_training - Step 15265: {'lr': 0.0004115518078864675, 'samples': 2931072, 'steps': 15265, 'loss/train': 0.4887765794992447} +01/27/2022 10:12:59 - INFO - codeparrot_training - Step 15266: {'lr': 0.0004115393203316946, 'samples': 2931264, 'steps': 15266, 'loss/train': 0.906529426574707} +01/27/2022 10:13:04 - INFO - codeparrot_training - Step 15267: {'lr': 0.00041152683208493855, 'samples': 2931456, 'steps': 15267, 'loss/train': 0.2713078334927559} +01/27/2022 10:13:07 - INFO - codeparrot_training - Step 15268: {'lr': 0.0004115143431462529, 'samples': 2931648, 'steps': 15268, 'loss/train': 1.2759371101856232} +01/27/2022 10:13:11 - INFO - codeparrot_training - Step 15269: {'lr': 0.000411501853515691, 'samples': 2931840, 'steps': 15269, 'loss/train': 0.742437869310379} +01/27/2022 10:13:14 - INFO - codeparrot_training - Step 15270: {'lr': 0.00041148936319330656, 'samples': 2932032, 'steps': 15270, 'loss/train': 0.757684975862503} +01/27/2022 10:13:17 - INFO - codeparrot_training - Step 15271: {'lr': 0.0004114768721791529, 'samples': 2932224, 'steps': 15271, 'loss/train': 0.9529096484184265} +01/27/2022 10:13:20 - INFO - codeparrot_training - Step 15272: {'lr': 0.00041146438047328347, 'samples': 2932416, 'steps': 15272, 'loss/train': 0.4093526601791382} +01/27/2022 10:13:23 - INFO - codeparrot_training - Step 15273: {'lr': 0.00041145188807575206, 'samples': 2932608, 'steps': 15273, 'loss/train': 0.8265180587768555} +01/27/2022 10:13:26 - INFO - codeparrot_training - Step 15274: {'lr': 0.000411439394986612, 'samples': 2932800, 'steps': 15274, 'loss/train': 0.7537222802639008} +01/27/2022 10:13:29 - INFO - codeparrot_training - Step 15275: {'lr': 0.00041142690120591686, 'samples': 2932992, 'steps': 15275, 'loss/train': 0.7074666023254395} +01/27/2022 10:13:34 - INFO - codeparrot_training - Step 15276: {'lr': 0.0004114144067337201, 'samples': 2933184, 'steps': 15276, 'loss/train': 0.6893979460000992} +01/27/2022 10:13:37 - INFO - codeparrot_training - Step 15277: {'lr': 0.0004114019115700752, 'samples': 2933376, 'steps': 15277, 'loss/train': 1.6662030816078186} +01/27/2022 10:13:40 - INFO - codeparrot_training - Step 15278: {'lr': 0.00041138941571503587, 'samples': 2933568, 'steps': 15278, 'loss/train': 0.8236050009727478} +01/27/2022 10:13:43 - INFO - codeparrot_training - Step 15279: {'lr': 0.0004113769191686555, 'samples': 2933760, 'steps': 15279, 'loss/train': 0.7639484703540802} +01/27/2022 10:13:46 - INFO - codeparrot_training - Step 15280: {'lr': 0.00041136442193098765, 'samples': 2933952, 'steps': 15280, 'loss/train': 1.1968939304351807} +01/27/2022 10:13:50 - INFO - codeparrot_training - Step 15281: {'lr': 0.00041135192400208585, 'samples': 2934144, 'steps': 15281, 'loss/train': 1.1819135248661041} +01/27/2022 10:13:53 - INFO - codeparrot_training - Step 15282: {'lr': 0.00041133942538200364, 'samples': 2934336, 'steps': 15282, 'loss/train': 0.8196035027503967} +01/27/2022 10:13:56 - INFO - codeparrot_training - Step 15283: {'lr': 0.0004113269260707946, 'samples': 2934528, 'steps': 15283, 'loss/train': 0.9080958366394043} +01/27/2022 10:14:00 - INFO - codeparrot_training - Step 15284: {'lr': 0.0004113144260685122, 'samples': 2934720, 'steps': 15284, 'loss/train': 0.6088594347238541} +01/27/2022 10:14:03 - INFO - codeparrot_training - Step 15285: {'lr': 0.00041130192537521, 'samples': 2934912, 'steps': 15285, 'loss/train': 0.5103038549423218} +01/27/2022 10:14:07 - INFO - codeparrot_training - Step 15286: {'lr': 0.0004112894239909416, 'samples': 2935104, 'steps': 15286, 'loss/train': 0.2525225132703781} +01/27/2022 10:14:10 - INFO - codeparrot_training - Step 15287: {'lr': 0.0004112769219157605, 'samples': 2935296, 'steps': 15287, 'loss/train': 0.39502689242362976} +01/27/2022 10:14:13 - INFO - codeparrot_training - Step 15288: {'lr': 0.00041126441914972036, 'samples': 2935488, 'steps': 15288, 'loss/train': 0.831643670797348} +01/27/2022 10:14:16 - INFO - codeparrot_training - Step 15289: {'lr': 0.00041125191569287456, 'samples': 2935680, 'steps': 15289, 'loss/train': 1.1580512523651123} +01/27/2022 10:14:19 - INFO - codeparrot_training - Step 15290: {'lr': 0.0004112394115452768, 'samples': 2935872, 'steps': 15290, 'loss/train': 0.6716718524694443} +01/27/2022 10:14:22 - INFO - codeparrot_training - Step 15291: {'lr': 0.00041122690670698054, 'samples': 2936064, 'steps': 15291, 'loss/train': 0.6267798095941544} +01/27/2022 10:14:25 - INFO - codeparrot_training - Step 15292: {'lr': 0.0004112144011780395, 'samples': 2936256, 'steps': 15292, 'loss/train': 0.7489550560712814} +01/27/2022 10:14:30 - INFO - codeparrot_training - Step 15293: {'lr': 0.00041120189495850713, 'samples': 2936448, 'steps': 15293, 'loss/train': 1.079516440629959} +01/27/2022 10:14:33 - INFO - codeparrot_training - Step 15294: {'lr': 0.000411189388048437, 'samples': 2936640, 'steps': 15294, 'loss/train': 0.9992784261703491} +01/27/2022 10:14:36 - INFO - codeparrot_training - Step 15295: {'lr': 0.0004111768804478827, 'samples': 2936832, 'steps': 15295, 'loss/train': 0.6946471184492111} +01/27/2022 10:14:39 - INFO - codeparrot_training - Step 15296: {'lr': 0.00041116437215689785, 'samples': 2937024, 'steps': 15296, 'loss/train': 0.9386417269706726} +01/27/2022 10:14:43 - INFO - codeparrot_training - Step 15297: {'lr': 0.000411151863175536, 'samples': 2937216, 'steps': 15297, 'loss/train': 0.9065860211849213} +01/27/2022 10:14:46 - INFO - codeparrot_training - Step 15298: {'lr': 0.00041113935350385074, 'samples': 2937408, 'steps': 15298, 'loss/train': 0.6236107349395752} +01/27/2022 10:14:49 - INFO - codeparrot_training - Step 15299: {'lr': 0.0004111268431418957, 'samples': 2937600, 'steps': 15299, 'loss/train': 0.8787169754505157} +01/27/2022 10:14:52 - INFO - codeparrot_training - Step 15300: {'lr': 0.0004111143320897244, 'samples': 2937792, 'steps': 15300, 'loss/train': 0.9310993552207947} +01/27/2022 10:14:55 - INFO - codeparrot_training - Step 15301: {'lr': 0.0004111018203473904, 'samples': 2937984, 'steps': 15301, 'loss/train': 0.8509276807308197} +01/27/2022 10:15:01 - INFO - codeparrot_training - Step 15302: {'lr': 0.0004110893079149474, 'samples': 2938176, 'steps': 15302, 'loss/train': 0.9827264249324799} +01/27/2022 10:15:04 - INFO - codeparrot_training - Step 15303: {'lr': 0.000411076794792449, 'samples': 2938368, 'steps': 15303, 'loss/train': 0.645241767168045} +01/27/2022 10:15:07 - INFO - codeparrot_training - Step 15304: {'lr': 0.0004110642809799487, 'samples': 2938560, 'steps': 15304, 'loss/train': 0.6606384068727493} +01/27/2022 10:15:10 - INFO - codeparrot_training - Step 15305: {'lr': 0.0004110517664775002, 'samples': 2938752, 'steps': 15305, 'loss/train': 1.1449845135211945} +01/27/2022 10:15:13 - INFO - codeparrot_training - Step 15306: {'lr': 0.00041103925128515705, 'samples': 2938944, 'steps': 15306, 'loss/train': 0.6412767469882965} +01/27/2022 10:15:17 - INFO - codeparrot_training - Step 15307: {'lr': 0.0004110267354029729, 'samples': 2939136, 'steps': 15307, 'loss/train': 0.7992390096187592} +01/27/2022 10:15:20 - INFO - codeparrot_training - Step 15308: {'lr': 0.0004110142188310013, 'samples': 2939328, 'steps': 15308, 'loss/train': 0.792874664068222} +01/27/2022 10:15:23 - INFO - codeparrot_training - Step 15309: {'lr': 0.00041100170156929596, 'samples': 2939520, 'steps': 15309, 'loss/train': 0.8168321549892426} +01/27/2022 10:15:26 - INFO - codeparrot_training - Step 15310: {'lr': 0.0004109891836179105, 'samples': 2939712, 'steps': 15310, 'loss/train': 0.7460162043571472} +01/27/2022 10:15:30 - INFO - codeparrot_training - Step 15311: {'lr': 0.0004109766649768984, 'samples': 2939904, 'steps': 15311, 'loss/train': 0.8948601186275482} +01/27/2022 10:15:34 - INFO - codeparrot_training - Step 15312: {'lr': 0.00041096414564631347, 'samples': 2940096, 'steps': 15312, 'loss/train': 0.9286184012889862} +01/27/2022 10:15:37 - INFO - codeparrot_training - Step 15313: {'lr': 0.00041095162562620915, 'samples': 2940288, 'steps': 15313, 'loss/train': 0.8153792023658752} +01/27/2022 10:15:40 - INFO - codeparrot_training - Step 15314: {'lr': 0.00041093910491663926, 'samples': 2940480, 'steps': 15314, 'loss/train': 1.0667567253112793} +01/27/2022 10:15:43 - INFO - codeparrot_training - Step 15315: {'lr': 0.0004109265835176573, 'samples': 2940672, 'steps': 15315, 'loss/train': 1.0862012207508087} +01/27/2022 10:15:46 - INFO - codeparrot_training - Step 15316: {'lr': 0.00041091406142931705, 'samples': 2940864, 'steps': 15316, 'loss/train': 0.6550306230783463} +01/27/2022 10:15:49 - INFO - codeparrot_training - Step 15317: {'lr': 0.00041090153865167196, 'samples': 2941056, 'steps': 15317, 'loss/train': 0.7640447616577148} +01/27/2022 10:15:52 - INFO - codeparrot_training - Step 15318: {'lr': 0.0004108890151847758, 'samples': 2941248, 'steps': 15318, 'loss/train': 0.6777294427156448} +01/27/2022 10:15:56 - INFO - codeparrot_training - Step 15319: {'lr': 0.0004108764910286822, 'samples': 2941440, 'steps': 15319, 'loss/train': 1.3649592697620392} +01/27/2022 10:16:00 - INFO - codeparrot_training - Step 15320: {'lr': 0.00041086396618344475, 'samples': 2941632, 'steps': 15320, 'loss/train': 0.6145138442516327} +01/27/2022 10:16:03 - INFO - codeparrot_training - Step 15321: {'lr': 0.0004108514406491172, 'samples': 2941824, 'steps': 15321, 'loss/train': 0.7728468775749207} +01/27/2022 10:16:06 - INFO - codeparrot_training - Step 15322: {'lr': 0.0004108389144257531, 'samples': 2942016, 'steps': 15322, 'loss/train': 1.0586694180965424} +01/27/2022 10:16:09 - INFO - codeparrot_training - Step 15323: {'lr': 0.0004108263875134062, 'samples': 2942208, 'steps': 15323, 'loss/train': 0.7949241399765015} +01/27/2022 10:16:13 - INFO - codeparrot_training - Step 15324: {'lr': 0.0004108138599121301, 'samples': 2942400, 'steps': 15324, 'loss/train': 0.5776965469121933} +01/27/2022 10:16:16 - INFO - codeparrot_training - Step 15325: {'lr': 0.00041080133162197855, 'samples': 2942592, 'steps': 15325, 'loss/train': 1.0449239015579224} +01/27/2022 10:16:19 - INFO - codeparrot_training - Step 15326: {'lr': 0.0004107888026430051, 'samples': 2942784, 'steps': 15326, 'loss/train': 1.0155149102210999} +01/27/2022 10:16:22 - INFO - codeparrot_training - Step 15327: {'lr': 0.0004107762729752635, 'samples': 2942976, 'steps': 15327, 'loss/train': 0.28637829422950745} +01/27/2022 10:16:25 - INFO - codeparrot_training - Step 15328: {'lr': 0.00041076374261880735, 'samples': 2943168, 'steps': 15328, 'loss/train': 0.9560295939445496} +01/27/2022 10:16:31 - INFO - codeparrot_training - Step 15329: {'lr': 0.0004107512115736904, 'samples': 2943360, 'steps': 15329, 'loss/train': 0.7067835330963135} +01/27/2022 10:16:34 - INFO - codeparrot_training - Step 15330: {'lr': 0.0004107386798399664, 'samples': 2943552, 'steps': 15330, 'loss/train': 0.7481270581483841} +01/27/2022 10:16:37 - INFO - codeparrot_training - Step 15331: {'lr': 0.00041072614741768877, 'samples': 2943744, 'steps': 15331, 'loss/train': 0.472157746553421} +01/27/2022 10:16:40 - INFO - codeparrot_training - Step 15332: {'lr': 0.00041071361430691143, 'samples': 2943936, 'steps': 15332, 'loss/train': 0.8282226026058197} +01/27/2022 10:16:43 - INFO - codeparrot_training - Step 15333: {'lr': 0.00041070108050768805, 'samples': 2944128, 'steps': 15333, 'loss/train': 5.481606602668762} +01/27/2022 10:16:47 - INFO - codeparrot_training - Step 15334: {'lr': 0.00041068854602007224, 'samples': 2944320, 'steps': 15334, 'loss/train': 0.7248976528644562} +01/27/2022 10:16:50 - INFO - codeparrot_training - Step 15335: {'lr': 0.0004106760108441177, 'samples': 2944512, 'steps': 15335, 'loss/train': 1.2078734636306763} +01/27/2022 10:16:53 - INFO - codeparrot_training - Step 15336: {'lr': 0.0004106634749798782, 'samples': 2944704, 'steps': 15336, 'loss/train': 1.0520188808441162} +01/27/2022 10:16:56 - INFO - codeparrot_training - Step 15337: {'lr': 0.0004106509384274073, 'samples': 2944896, 'steps': 15337, 'loss/train': 0.6389463096857071} +01/27/2022 10:17:00 - INFO - codeparrot_training - Step 15338: {'lr': 0.0004106384011867589, 'samples': 2945088, 'steps': 15338, 'loss/train': 0.5491106361150742} +01/27/2022 10:17:04 - INFO - codeparrot_training - Step 15339: {'lr': 0.00041062586325798654, 'samples': 2945280, 'steps': 15339, 'loss/train': 1.0565254390239716} +01/27/2022 10:17:07 - INFO - codeparrot_training - Step 15340: {'lr': 0.000410613324641144, 'samples': 2945472, 'steps': 15340, 'loss/train': 0.7047126889228821} +01/27/2022 10:17:10 - INFO - codeparrot_training - Step 15341: {'lr': 0.000410600785336285, 'samples': 2945664, 'steps': 15341, 'loss/train': 0.8681658804416656} +01/27/2022 10:17:13 - INFO - codeparrot_training - Step 15342: {'lr': 0.0004105882453434632, 'samples': 2945856, 'steps': 15342, 'loss/train': 0.7791650891304016} +01/27/2022 10:17:16 - INFO - codeparrot_training - Step 15343: {'lr': 0.0004105757046627323, 'samples': 2946048, 'steps': 15343, 'loss/train': 0.44820772111415863} +01/27/2022 10:17:19 - INFO - codeparrot_training - Step 15344: {'lr': 0.00041056316329414613, 'samples': 2946240, 'steps': 15344, 'loss/train': 0.3611082136631012} +01/27/2022 10:17:22 - INFO - codeparrot_training - Step 15345: {'lr': 0.0004105506212377583, 'samples': 2946432, 'steps': 15345, 'loss/train': 1.0265438854694366} +01/27/2022 10:17:28 - INFO - codeparrot_training - Step 15346: {'lr': 0.0004105380784936227, 'samples': 2946624, 'steps': 15346, 'loss/train': 1.035628080368042} +01/27/2022 10:17:31 - INFO - codeparrot_training - Step 15347: {'lr': 0.0004105255350617928, 'samples': 2946816, 'steps': 15347, 'loss/train': 0.7671354711055756} +01/27/2022 10:17:34 - INFO - codeparrot_training - Step 15348: {'lr': 0.0004105129909423226, 'samples': 2947008, 'steps': 15348, 'loss/train': 0.9211370944976807} +01/27/2022 10:17:37 - INFO - codeparrot_training - Step 15349: {'lr': 0.0004105004461352657, 'samples': 2947200, 'steps': 15349, 'loss/train': 1.0758951902389526} +01/27/2022 10:17:40 - INFO - codeparrot_training - Step 15350: {'lr': 0.00041048790064067577, 'samples': 2947392, 'steps': 15350, 'loss/train': 1.1553095877170563} +01/27/2022 10:17:43 - INFO - codeparrot_training - Step 15351: {'lr': 0.0004104753544586067, 'samples': 2947584, 'steps': 15351, 'loss/train': 0.8947143852710724} +01/27/2022 10:17:46 - INFO - codeparrot_training - Step 15352: {'lr': 0.0004104628075891121, 'samples': 2947776, 'steps': 15352, 'loss/train': 0.7557309865951538} +01/27/2022 10:17:50 - INFO - codeparrot_training - Step 15353: {'lr': 0.00041045026003224593, 'samples': 2947968, 'steps': 15353, 'loss/train': 1.0776202976703644} +01/27/2022 10:17:53 - INFO - codeparrot_training - Step 15354: {'lr': 0.00041043771178806164, 'samples': 2948160, 'steps': 15354, 'loss/train': 0.5889621376991272} +01/27/2022 10:17:57 - INFO - codeparrot_training - Step 15355: {'lr': 0.00041042516285661325, 'samples': 2948352, 'steps': 15355, 'loss/train': 0.6216762810945511} +01/27/2022 10:18:00 - INFO - codeparrot_training - Step 15356: {'lr': 0.00041041261323795437, 'samples': 2948544, 'steps': 15356, 'loss/train': 0.7122577428817749} +01/27/2022 10:18:03 - INFO - codeparrot_training - Step 15357: {'lr': 0.00041040006293213883, 'samples': 2948736, 'steps': 15357, 'loss/train': 0.7970263659954071} +01/27/2022 10:18:07 - INFO - codeparrot_training - Step 15358: {'lr': 0.0004103875119392203, 'samples': 2948928, 'steps': 15358, 'loss/train': 1.0583327114582062} +01/27/2022 10:18:10 - INFO - codeparrot_training - Step 15359: {'lr': 0.00041037496025925256, 'samples': 2949120, 'steps': 15359, 'loss/train': 0.7347071617841721} +01/27/2022 10:18:13 - INFO - codeparrot_training - Step 15360: {'lr': 0.0004103624078922895, 'samples': 2949312, 'steps': 15360, 'loss/train': 1.1761009991168976} +01/27/2022 10:18:16 - INFO - codeparrot_training - Step 15361: {'lr': 0.0004103498548383847, 'samples': 2949504, 'steps': 15361, 'loss/train': 1.86068594455719} +01/27/2022 10:18:19 - INFO - codeparrot_training - Step 15362: {'lr': 0.00041033730109759216, 'samples': 2949696, 'steps': 15362, 'loss/train': 0.7421299070119858} +01/27/2022 10:18:22 - INFO - codeparrot_training - Step 15363: {'lr': 0.00041032474666996544, 'samples': 2949888, 'steps': 15363, 'loss/train': 0.8323457837104797} +01/27/2022 10:18:27 - INFO - codeparrot_training - Step 15364: {'lr': 0.0004103121915555585, 'samples': 2950080, 'steps': 15364, 'loss/train': 0.8781857192516327} +01/27/2022 10:18:30 - INFO - codeparrot_training - Step 15365: {'lr': 0.00041029963575442494, 'samples': 2950272, 'steps': 15365, 'loss/train': 0.6962047219276428} +01/27/2022 10:18:33 - INFO - codeparrot_training - Step 15366: {'lr': 0.0004102870792666187, 'samples': 2950464, 'steps': 15366, 'loss/train': 0.7535512447357178} +01/27/2022 10:18:36 - INFO - codeparrot_training - Step 15367: {'lr': 0.0004102745220921935, 'samples': 2950656, 'steps': 15367, 'loss/train': 1.1577123999595642} +01/27/2022 10:18:39 - INFO - codeparrot_training - Step 15368: {'lr': 0.0004102619642312031, 'samples': 2950848, 'steps': 15368, 'loss/train': 1.0223055481910706} +01/27/2022 10:18:43 - INFO - codeparrot_training - Step 15369: {'lr': 0.0004102494056837014, 'samples': 2951040, 'steps': 15369, 'loss/train': 0.6968680769205093} +01/27/2022 10:18:46 - INFO - codeparrot_training - Step 15370: {'lr': 0.00041023684644974213, 'samples': 2951232, 'steps': 15370, 'loss/train': 0.3338806629180908} +01/27/2022 10:18:49 - INFO - codeparrot_training - Step 15371: {'lr': 0.00041022428652937905, 'samples': 2951424, 'steps': 15371, 'loss/train': 0.6201076805591583} +01/27/2022 10:18:52 - INFO - codeparrot_training - Step 15372: {'lr': 0.000410211725922666, 'samples': 2951616, 'steps': 15372, 'loss/train': 0.5547677725553513} +01/27/2022 10:18:58 - INFO - codeparrot_training - Step 15373: {'lr': 0.00041019916462965684, 'samples': 2951808, 'steps': 15373, 'loss/train': 0.3629724830389023} +01/27/2022 10:19:01 - INFO - codeparrot_training - Step 15374: {'lr': 0.0004101866026504053, 'samples': 2952000, 'steps': 15374, 'loss/train': 0.852946400642395} +01/27/2022 10:19:04 - INFO - codeparrot_training - Step 15375: {'lr': 0.00041017403998496523, 'samples': 2952192, 'steps': 15375, 'loss/train': 0.7558294236660004} +01/27/2022 10:19:07 - INFO - codeparrot_training - Step 15376: {'lr': 0.0004101614766333904, 'samples': 2952384, 'steps': 15376, 'loss/train': 0.872931182384491} +01/27/2022 10:19:10 - INFO - codeparrot_training - Step 15377: {'lr': 0.0004101489125957347, 'samples': 2952576, 'steps': 15377, 'loss/train': 0.8523485362529755} +01/27/2022 10:19:13 - INFO - codeparrot_training - Step 15378: {'lr': 0.0004101363478720519, 'samples': 2952768, 'steps': 15378, 'loss/train': 1.1727977693080902} +01/27/2022 10:19:16 - INFO - codeparrot_training - Step 15379: {'lr': 0.0004101237824623958, 'samples': 2952960, 'steps': 15379, 'loss/train': 1.3045495748519897} +01/27/2022 10:19:20 - INFO - codeparrot_training - Step 15380: {'lr': 0.00041011121636682024, 'samples': 2953152, 'steps': 15380, 'loss/train': 0.8636935651302338} +01/27/2022 10:19:23 - INFO - codeparrot_training - Step 15381: {'lr': 0.0004100986495853791, 'samples': 2953344, 'steps': 15381, 'loss/train': 0.8260681629180908} +01/27/2022 10:19:27 - INFO - codeparrot_training - Step 15382: {'lr': 0.00041008608211812625, 'samples': 2953536, 'steps': 15382, 'loss/train': 0.8720628619194031} +01/27/2022 10:19:31 - INFO - codeparrot_training - Step 15383: {'lr': 0.00041007351396511537, 'samples': 2953728, 'steps': 15383, 'loss/train': 0.48479560017585754} +01/27/2022 10:19:34 - INFO - codeparrot_training - Step 15384: {'lr': 0.00041006094512640044, 'samples': 2953920, 'steps': 15384, 'loss/train': 0.9024207293987274} +01/27/2022 10:19:37 - INFO - codeparrot_training - Step 15385: {'lr': 0.00041004837560203525, 'samples': 2954112, 'steps': 15385, 'loss/train': 0.6653470695018768} +01/27/2022 10:19:40 - INFO - codeparrot_training - Step 15386: {'lr': 0.0004100358053920736, 'samples': 2954304, 'steps': 15386, 'loss/train': 0.4877624362707138} +01/27/2022 10:19:43 - INFO - codeparrot_training - Step 15387: {'lr': 0.00041002323449656943, 'samples': 2954496, 'steps': 15387, 'loss/train': 0.532447099685669} +01/27/2022 10:19:46 - INFO - codeparrot_training - Step 15388: {'lr': 0.00041001066291557653, 'samples': 2954688, 'steps': 15388, 'loss/train': 0.7097757160663605} +01/27/2022 10:19:49 - INFO - codeparrot_training - Step 15389: {'lr': 0.0004099980906491487, 'samples': 2954880, 'steps': 15389, 'loss/train': 0.6709431409835815} +01/27/2022 10:19:54 - INFO - codeparrot_training - Step 15390: {'lr': 0.0004099855176973399, 'samples': 2955072, 'steps': 15390, 'loss/train': 1.2830601632595062} +01/27/2022 10:19:57 - INFO - codeparrot_training - Step 15391: {'lr': 0.0004099729440602039, 'samples': 2955264, 'steps': 15391, 'loss/train': 0.842666745185852} +01/27/2022 10:20:00 - INFO - codeparrot_training - Step 15392: {'lr': 0.0004099603697377946, 'samples': 2955456, 'steps': 15392, 'loss/train': 0.3614940941333771} +01/27/2022 10:20:03 - INFO - codeparrot_training - Step 15393: {'lr': 0.000409947794730166, 'samples': 2955648, 'steps': 15393, 'loss/train': 0.3422804921865463} +01/27/2022 10:20:07 - INFO - codeparrot_training - Step 15394: {'lr': 0.0004099352190373716, 'samples': 2955840, 'steps': 15394, 'loss/train': 1.1397319436073303} +01/27/2022 10:20:10 - INFO - codeparrot_training - Step 15395: {'lr': 0.0004099226426594657, 'samples': 2956032, 'steps': 15395, 'loss/train': 0.5593622177839279} +01/27/2022 10:20:13 - INFO - codeparrot_training - Step 15396: {'lr': 0.0004099100655965019, 'samples': 2956224, 'steps': 15396, 'loss/train': 0.3328387588262558} +01/27/2022 10:20:16 - INFO - codeparrot_training - Step 15397: {'lr': 0.0004098974878485342, 'samples': 2956416, 'steps': 15397, 'loss/train': 1.0343619883060455} +01/27/2022 10:20:19 - INFO - codeparrot_training - Step 15398: {'lr': 0.0004098849094156164, 'samples': 2956608, 'steps': 15398, 'loss/train': 0.26610899716615677} +01/27/2022 10:20:24 - INFO - codeparrot_training - Step 15399: {'lr': 0.0004098723302978025, 'samples': 2956800, 'steps': 15399, 'loss/train': 1.0840213894844055} +01/27/2022 10:20:27 - INFO - codeparrot_training - Step 15400: {'lr': 0.00040985975049514617, 'samples': 2956992, 'steps': 15400, 'loss/train': 0.974641889333725} +01/27/2022 10:20:30 - INFO - codeparrot_training - Step 15401: {'lr': 0.00040984717000770157, 'samples': 2957184, 'steps': 15401, 'loss/train': 0.5797430723905563} +01/27/2022 10:20:33 - INFO - codeparrot_training - Step 15402: {'lr': 0.00040983458883552237, 'samples': 2957376, 'steps': 15402, 'loss/train': 1.0707774460315704} +01/27/2022 10:20:36 - INFO - codeparrot_training - Step 15403: {'lr': 0.00040982200697866256, 'samples': 2957568, 'steps': 15403, 'loss/train': 0.7000829726457596} +01/27/2022 10:20:39 - INFO - codeparrot_training - Step 15404: {'lr': 0.00040980942443717596, 'samples': 2957760, 'steps': 15404, 'loss/train': 0.9079862236976624} +01/27/2022 10:20:42 - INFO - codeparrot_training - Step 15405: {'lr': 0.0004097968412111166, 'samples': 2957952, 'steps': 15405, 'loss/train': 0.5042577087879181} +01/27/2022 10:20:46 - INFO - codeparrot_training - Step 15406: {'lr': 0.0004097842573005383, 'samples': 2958144, 'steps': 15406, 'loss/train': 1.1329558789730072} +01/27/2022 10:20:49 - INFO - codeparrot_training - Step 15407: {'lr': 0.000409771672705495, 'samples': 2958336, 'steps': 15407, 'loss/train': 0.7826457917690277} +01/27/2022 10:20:54 - INFO - codeparrot_training - Step 15408: {'lr': 0.0004097590874260405, 'samples': 2958528, 'steps': 15408, 'loss/train': 0.765110582113266} +01/27/2022 10:20:57 - INFO - codeparrot_training - Step 15409: {'lr': 0.0004097465014622289, 'samples': 2958720, 'steps': 15409, 'loss/train': 0.8616582155227661} +01/27/2022 10:21:00 - INFO - codeparrot_training - Step 15410: {'lr': 0.00040973391481411396, 'samples': 2958912, 'steps': 15410, 'loss/train': 0.803503543138504} +01/27/2022 10:21:03 - INFO - codeparrot_training - Step 15411: {'lr': 0.00040972132748174966, 'samples': 2959104, 'steps': 15411, 'loss/train': 0.6397913843393326} +01/27/2022 10:21:07 - INFO - codeparrot_training - Step 15412: {'lr': 0.00040970873946518993, 'samples': 2959296, 'steps': 15412, 'loss/train': 0.3910711705684662} +01/27/2022 10:21:10 - INFO - codeparrot_training - Step 15413: {'lr': 0.00040969615076448865, 'samples': 2959488, 'steps': 15413, 'loss/train': 0.5715521425008774} +01/27/2022 10:21:13 - INFO - codeparrot_training - Step 15414: {'lr': 0.0004096835613796998, 'samples': 2959680, 'steps': 15414, 'loss/train': 0.8066655993461609} +01/27/2022 10:21:16 - INFO - codeparrot_training - Step 15415: {'lr': 0.00040967097131087727, 'samples': 2959872, 'steps': 15415, 'loss/train': 0.832080066204071} +01/27/2022 10:21:19 - INFO - codeparrot_training - Step 15416: {'lr': 0.00040965838055807493, 'samples': 2960064, 'steps': 15416, 'loss/train': 1.0196921825408936} +01/27/2022 10:21:24 - INFO - codeparrot_training - Step 15417: {'lr': 0.00040964578912134687, 'samples': 2960256, 'steps': 15417, 'loss/train': 0.29385410249233246} +01/27/2022 10:21:27 - INFO - codeparrot_training - Step 15418: {'lr': 0.00040963319700074684, 'samples': 2960448, 'steps': 15418, 'loss/train': 0.7163080126047134} +01/27/2022 10:21:30 - INFO - codeparrot_training - Step 15419: {'lr': 0.00040962060419632906, 'samples': 2960640, 'steps': 15419, 'loss/train': 0.8143856227397919} +01/27/2022 10:21:33 - INFO - codeparrot_training - Step 15420: {'lr': 0.00040960801070814715, 'samples': 2960832, 'steps': 15420, 'loss/train': 0.7456417679786682} +01/27/2022 10:21:36 - INFO - codeparrot_training - Step 15421: {'lr': 0.00040959541653625526, 'samples': 2961024, 'steps': 15421, 'loss/train': 1.076731413602829} +01/27/2022 10:21:39 - INFO - codeparrot_training - Step 15422: {'lr': 0.0004095828216807073, 'samples': 2961216, 'steps': 15422, 'loss/train': 1.050427794456482} +01/27/2022 10:21:43 - INFO - codeparrot_training - Step 15423: {'lr': 0.00040957022614155714, 'samples': 2961408, 'steps': 15423, 'loss/train': 0.9452986121177673} +01/27/2022 10:21:46 - INFO - codeparrot_training - Step 15424: {'lr': 0.0004095576299188589, 'samples': 2961600, 'steps': 15424, 'loss/train': 0.7747955024242401} +01/27/2022 10:21:49 - INFO - codeparrot_training - Step 15425: {'lr': 0.0004095450330126663, 'samples': 2961792, 'steps': 15425, 'loss/train': 0.5942653566598892} +01/27/2022 10:21:54 - INFO - codeparrot_training - Step 15426: {'lr': 0.0004095324354230335, 'samples': 2961984, 'steps': 15426, 'loss/train': 0.6088310480117798} +01/27/2022 10:21:57 - INFO - codeparrot_training - Step 15427: {'lr': 0.0004095198371500145, 'samples': 2962176, 'steps': 15427, 'loss/train': 0.8906205296516418} +01/27/2022 10:22:00 - INFO - codeparrot_training - Step 15428: {'lr': 0.00040950723819366307, 'samples': 2962368, 'steps': 15428, 'loss/train': 0.8517653346061707} +01/27/2022 10:22:03 - INFO - codeparrot_training - Step 15429: {'lr': 0.00040949463855403326, 'samples': 2962560, 'steps': 15429, 'loss/train': 0.7726511657238007} +01/27/2022 10:22:06 - INFO - codeparrot_training - Step 15430: {'lr': 0.00040948203823117915, 'samples': 2962752, 'steps': 15430, 'loss/train': 1.1972092688083649} +01/27/2022 10:22:10 - INFO - codeparrot_training - Step 15431: {'lr': 0.00040946943722515455, 'samples': 2962944, 'steps': 15431, 'loss/train': 0.569014772772789} +01/27/2022 10:22:13 - INFO - codeparrot_training - Step 15432: {'lr': 0.0004094568355360135, 'samples': 2963136, 'steps': 15432, 'loss/train': 0.9697061777114868} +01/27/2022 10:22:16 - INFO - codeparrot_training - Step 15433: {'lr': 0.00040944423316381006, 'samples': 2963328, 'steps': 15433, 'loss/train': 0.28715329617261887} +01/27/2022 10:22:20 - INFO - codeparrot_training - Step 15434: {'lr': 0.0004094316301085982, 'samples': 2963520, 'steps': 15434, 'loss/train': 0.8934229016304016} +01/27/2022 10:22:23 - INFO - codeparrot_training - Step 15435: {'lr': 0.00040941902637043183, 'samples': 2963712, 'steps': 15435, 'loss/train': 0.6711100190877914} +01/27/2022 10:22:27 - INFO - codeparrot_training - Step 15436: {'lr': 0.00040940642194936495, 'samples': 2963904, 'steps': 15436, 'loss/train': 0.7659207880496979} +01/27/2022 10:22:30 - INFO - codeparrot_training - Step 15437: {'lr': 0.0004093938168454515, 'samples': 2964096, 'steps': 15437, 'loss/train': 0.9133318662643433} +01/27/2022 10:22:33 - INFO - codeparrot_training - Step 15438: {'lr': 0.00040938121105874573, 'samples': 2964288, 'steps': 15438, 'loss/train': 0.9779216051101685} +01/27/2022 10:22:36 - INFO - codeparrot_training - Step 15439: {'lr': 0.0004093686045893013, 'samples': 2964480, 'steps': 15439, 'loss/train': 0.22707512229681015} +01/27/2022 10:22:39 - INFO - codeparrot_training - Step 15440: {'lr': 0.00040935599743717243, 'samples': 2964672, 'steps': 15440, 'loss/train': 0.9135463535785675} +01/27/2022 10:22:42 - INFO - codeparrot_training - Step 15441: {'lr': 0.00040934338960241305, 'samples': 2964864, 'steps': 15441, 'loss/train': 0.6397002339363098} +01/27/2022 10:22:45 - INFO - codeparrot_training - Step 15442: {'lr': 0.00040933078108507727, 'samples': 2965056, 'steps': 15442, 'loss/train': 1.0284152626991272} +01/27/2022 10:22:50 - INFO - codeparrot_training - Step 15443: {'lr': 0.00040931817188521894, 'samples': 2965248, 'steps': 15443, 'loss/train': 0.9298297762870789} +01/27/2022 10:22:53 - INFO - codeparrot_training - Step 15444: {'lr': 0.00040930556200289214, 'samples': 2965440, 'steps': 15444, 'loss/train': 0.9625587165355682} +01/27/2022 10:22:56 - INFO - codeparrot_training - Step 15445: {'lr': 0.00040929295143815093, 'samples': 2965632, 'steps': 15445, 'loss/train': 0.1810937449336052} +01/27/2022 10:22:59 - INFO - codeparrot_training - Step 15446: {'lr': 0.0004092803401910493, 'samples': 2965824, 'steps': 15446, 'loss/train': 1.0016240179538727} +01/27/2022 10:23:02 - INFO - codeparrot_training - Step 15447: {'lr': 0.00040926772826164126, 'samples': 2966016, 'steps': 15447, 'loss/train': 1.3628455102443695} +01/27/2022 10:23:06 - INFO - codeparrot_training - Step 15448: {'lr': 0.0004092551156499809, 'samples': 2966208, 'steps': 15448, 'loss/train': 0.4611998051404953} +01/27/2022 10:23:09 - INFO - codeparrot_training - Step 15449: {'lr': 0.000409242502356122, 'samples': 2966400, 'steps': 15449, 'loss/train': 0.7133510112762451} +01/27/2022 10:23:12 - INFO - codeparrot_training - Step 15450: {'lr': 0.000409229888380119, 'samples': 2966592, 'steps': 15450, 'loss/train': 0.8699155747890472} +01/27/2022 10:23:15 - INFO - codeparrot_training - Step 15451: {'lr': 0.00040921727372202565, 'samples': 2966784, 'steps': 15451, 'loss/train': 0.39516958594322205} +01/27/2022 10:23:20 - INFO - codeparrot_training - Step 15452: {'lr': 0.000409204658381896, 'samples': 2966976, 'steps': 15452, 'loss/train': 0.8630495667457581} +01/27/2022 10:23:23 - INFO - codeparrot_training - Step 15453: {'lr': 0.00040919204235978425, 'samples': 2967168, 'steps': 15453, 'loss/train': 0.6643611788749695} +01/27/2022 10:23:26 - INFO - codeparrot_training - Step 15454: {'lr': 0.0004091794256557443, 'samples': 2967360, 'steps': 15454, 'loss/train': 1.0685891211032867} +01/27/2022 10:23:30 - INFO - codeparrot_training - Step 15455: {'lr': 0.00040916680826983017, 'samples': 2967552, 'steps': 15455, 'loss/train': 0.42240411043167114} +01/27/2022 10:23:33 - INFO - codeparrot_training - Step 15456: {'lr': 0.00040915419020209605, 'samples': 2967744, 'steps': 15456, 'loss/train': 0.2999411076307297} +01/27/2022 10:23:36 - INFO - codeparrot_training - Step 15457: {'lr': 0.0004091415714525959, 'samples': 2967936, 'steps': 15457, 'loss/train': 0.992893785238266} +01/27/2022 10:23:39 - INFO - codeparrot_training - Step 15458: {'lr': 0.0004091289520213838, 'samples': 2968128, 'steps': 15458, 'loss/train': 0.8570938110351562} +01/27/2022 10:23:42 - INFO - codeparrot_training - Step 15459: {'lr': 0.0004091163319085137, 'samples': 2968320, 'steps': 15459, 'loss/train': 0.9159821569919586} +01/27/2022 10:23:45 - INFO - codeparrot_training - Step 15460: {'lr': 0.0004091037111140399, 'samples': 2968512, 'steps': 15460, 'loss/train': 0.6374292522668839} +01/27/2022 10:23:50 - INFO - codeparrot_training - Step 15461: {'lr': 0.00040909108963801624, 'samples': 2968704, 'steps': 15461, 'loss/train': 0.33885468542575836} +01/27/2022 10:23:53 - INFO - codeparrot_training - Step 15462: {'lr': 0.0004090784674804969, 'samples': 2968896, 'steps': 15462, 'loss/train': 0.8767612874507904} +01/27/2022 10:23:56 - INFO - codeparrot_training - Step 15463: {'lr': 0.0004090658446415359, 'samples': 2969088, 'steps': 15463, 'loss/train': 1.3293474912643433} +01/27/2022 10:23:59 - INFO - codeparrot_training - Step 15464: {'lr': 0.0004090532211211874, 'samples': 2969280, 'steps': 15464, 'loss/train': 0.4809568226337433} +01/27/2022 10:24:02 - INFO - codeparrot_training - Step 15465: {'lr': 0.0004090405969195053, 'samples': 2969472, 'steps': 15465, 'loss/train': 0.7625145614147186} +01/27/2022 10:24:05 - INFO - codeparrot_training - Step 15466: {'lr': 0.0004090279720365438, 'samples': 2969664, 'steps': 15466, 'loss/train': 0.42583778500556946} +01/27/2022 10:24:09 - INFO - codeparrot_training - Step 15467: {'lr': 0.00040901534647235703, 'samples': 2969856, 'steps': 15467, 'loss/train': 0.6243401169776917} +01/27/2022 10:24:12 - INFO - codeparrot_training - Step 15468: {'lr': 0.00040900272022699897, 'samples': 2970048, 'steps': 15468, 'loss/train': 1.0161330699920654} +01/27/2022 10:24:15 - INFO - codeparrot_training - Step 15469: {'lr': 0.00040899009330052375, 'samples': 2970240, 'steps': 15469, 'loss/train': 0.6883056610822678} +01/27/2022 10:24:20 - INFO - codeparrot_training - Step 15470: {'lr': 0.00040897746569298546, 'samples': 2970432, 'steps': 15470, 'loss/train': 0.5604508370161057} +01/27/2022 10:24:23 - INFO - codeparrot_training - Step 15471: {'lr': 0.0004089648374044382, 'samples': 2970624, 'steps': 15471, 'loss/train': 0.35868293792009354} +01/27/2022 10:24:26 - INFO - codeparrot_training - Step 15472: {'lr': 0.000408952208434936, 'samples': 2970816, 'steps': 15472, 'loss/train': 1.0780014395713806} +01/27/2022 10:24:29 - INFO - codeparrot_training - Step 15473: {'lr': 0.00040893957878453314, 'samples': 2971008, 'steps': 15473, 'loss/train': 1.3253121972084045} +01/27/2022 10:24:32 - INFO - codeparrot_training - Step 15474: {'lr': 0.0004089269484532834, 'samples': 2971200, 'steps': 15474, 'loss/train': 0.974394679069519} +01/27/2022 10:24:36 - INFO - codeparrot_training - Step 15475: {'lr': 0.00040891431744124123, 'samples': 2971392, 'steps': 15475, 'loss/train': 0.9289584159851074} +01/27/2022 10:24:39 - INFO - codeparrot_training - Step 15476: {'lr': 0.00040890168574846055, 'samples': 2971584, 'steps': 15476, 'loss/train': 0.7677411139011383} +01/27/2022 10:24:42 - INFO - codeparrot_training - Step 15477: {'lr': 0.0004088890533749955, 'samples': 2971776, 'steps': 15477, 'loss/train': 0.4670577943325043} +01/27/2022 10:24:46 - INFO - codeparrot_training - Step 15478: {'lr': 0.0004088764203209002, 'samples': 2971968, 'steps': 15478, 'loss/train': 1.1226298213005066} +01/27/2022 10:24:49 - INFO - codeparrot_training - Step 15479: {'lr': 0.0004088637865862287, 'samples': 2972160, 'steps': 15479, 'loss/train': 0.2535514086484909} +01/27/2022 10:24:53 - INFO - codeparrot_training - Step 15480: {'lr': 0.0004088511521710352, 'samples': 2972352, 'steps': 15480, 'loss/train': 0.6481539756059647} +01/27/2022 10:24:56 - INFO - codeparrot_training - Step 15481: {'lr': 0.0004088385170753739, 'samples': 2972544, 'steps': 15481, 'loss/train': 0.8312601149082184} +01/27/2022 10:24:59 - INFO - codeparrot_training - Step 15482: {'lr': 0.00040882588129929876, 'samples': 2972736, 'steps': 15482, 'loss/train': 0.7791342437267303} +01/27/2022 10:25:02 - INFO - codeparrot_training - Step 15483: {'lr': 0.000408813244842864, 'samples': 2972928, 'steps': 15483, 'loss/train': 0.8003999590873718} +01/27/2022 10:25:05 - INFO - codeparrot_training - Step 15484: {'lr': 0.0004088006077061237, 'samples': 2973120, 'steps': 15484, 'loss/train': 0.32488737255334854} +01/27/2022 10:25:08 - INFO - codeparrot_training - Step 15485: {'lr': 0.00040878796988913204, 'samples': 2973312, 'steps': 15485, 'loss/train': 0.9065728783607483} +01/27/2022 10:25:11 - INFO - codeparrot_training - Step 15486: {'lr': 0.00040877533139194313, 'samples': 2973504, 'steps': 15486, 'loss/train': 0.6098978072404861} +01/27/2022 10:25:16 - INFO - codeparrot_training - Step 15487: {'lr': 0.00040876269221461117, 'samples': 2973696, 'steps': 15487, 'loss/train': 0.9121798574924469} +01/27/2022 10:25:19 - INFO - codeparrot_training - Step 15488: {'lr': 0.0004087500523571902, 'samples': 2973888, 'steps': 15488, 'loss/train': 0.7558278143405914} +01/27/2022 10:25:22 - INFO - codeparrot_training - Step 15489: {'lr': 0.0004087374118197344, 'samples': 2974080, 'steps': 15489, 'loss/train': 1.1846924722194672} +01/27/2022 10:25:25 - INFO - codeparrot_training - Step 15490: {'lr': 0.00040872477060229797, 'samples': 2974272, 'steps': 15490, 'loss/train': 0.8027832806110382} +01/27/2022 10:25:29 - INFO - codeparrot_training - Step 15491: {'lr': 0.00040871212870493504, 'samples': 2974464, 'steps': 15491, 'loss/train': 0.6275765597820282} +01/27/2022 10:25:32 - INFO - codeparrot_training - Step 15492: {'lr': 0.0004086994861276996, 'samples': 2974656, 'steps': 15492, 'loss/train': 0.8696827590465546} +01/27/2022 10:25:35 - INFO - codeparrot_training - Step 15493: {'lr': 0.00040868684287064617, 'samples': 2974848, 'steps': 15493, 'loss/train': 0.4872148185968399} +01/27/2022 10:25:38 - INFO - codeparrot_training - Step 15494: {'lr': 0.0004086741989338285, 'samples': 2975040, 'steps': 15494, 'loss/train': 1.120416909456253} +01/27/2022 10:25:41 - INFO - codeparrot_training - Step 15495: {'lr': 0.0004086615543173011, 'samples': 2975232, 'steps': 15495, 'loss/train': 0.7007139176130295} +01/27/2022 10:25:46 - INFO - codeparrot_training - Step 15496: {'lr': 0.0004086489090211178, 'samples': 2975424, 'steps': 15496, 'loss/train': 0.9246970117092133} +01/27/2022 10:25:49 - INFO - codeparrot_training - Step 15497: {'lr': 0.00040863626304533316, 'samples': 2975616, 'steps': 15497, 'loss/train': 0.8666257560253143} +01/27/2022 10:25:52 - INFO - codeparrot_training - Step 15498: {'lr': 0.000408623616390001, 'samples': 2975808, 'steps': 15498, 'loss/train': 0.6772508472204208} +01/27/2022 10:25:55 - INFO - codeparrot_training - Step 15499: {'lr': 0.00040861096905517574, 'samples': 2976000, 'steps': 15499, 'loss/train': 1.1238928735256195} +01/27/2022 10:25:58 - INFO - codeparrot_training - Step 15500: {'lr': 0.0004085983210409114, 'samples': 2976192, 'steps': 15500, 'loss/train': 1.052830070257187} +01/27/2022 10:26:01 - INFO - codeparrot_training - Step 15501: {'lr': 0.00040858567234726217, 'samples': 2976384, 'steps': 15501, 'loss/train': 0.8793705403804779} +01/27/2022 10:26:05 - INFO - codeparrot_training - Step 15502: {'lr': 0.00040857302297428233, 'samples': 2976576, 'steps': 15502, 'loss/train': 0.9307697117328644} +01/27/2022 10:26:08 - INFO - codeparrot_training - Step 15503: {'lr': 0.000408560372922026, 'samples': 2976768, 'steps': 15503, 'loss/train': 0.7388522028923035} +01/27/2022 10:26:11 - INFO - codeparrot_training - Step 15504: {'lr': 0.00040854772219054737, 'samples': 2976960, 'steps': 15504, 'loss/train': 0.27265483886003494} +01/27/2022 10:26:16 - INFO - codeparrot_training - Step 15505: {'lr': 0.00040853507077990073, 'samples': 2977152, 'steps': 15505, 'loss/train': 0.4460193067789078} +01/27/2022 10:26:20 - INFO - codeparrot_training - Step 15506: {'lr': 0.00040852241869014004, 'samples': 2977344, 'steps': 15506, 'loss/train': 1.0507761240005493} +01/27/2022 10:26:23 - INFO - codeparrot_training - Step 15507: {'lr': 0.00040850976592131974, 'samples': 2977536, 'steps': 15507, 'loss/train': 1.1300164461135864} +01/27/2022 10:26:26 - INFO - codeparrot_training - Step 15508: {'lr': 0.0004084971124734939, 'samples': 2977728, 'steps': 15508, 'loss/train': 0.7711881101131439} +01/27/2022 10:26:29 - INFO - codeparrot_training - Step 15509: {'lr': 0.0004084844583467168, 'samples': 2977920, 'steps': 15509, 'loss/train': 0.37747304141521454} +01/27/2022 10:26:32 - INFO - codeparrot_training - Step 15510: {'lr': 0.00040847180354104256, 'samples': 2978112, 'steps': 15510, 'loss/train': 0.9257469177246094} +01/27/2022 10:26:35 - INFO - codeparrot_training - Step 15511: {'lr': 0.00040845914805652544, 'samples': 2978304, 'steps': 15511, 'loss/train': 0.8829046189785004} +01/27/2022 10:26:38 - INFO - codeparrot_training - Step 15512: {'lr': 0.0004084464918932197, 'samples': 2978496, 'steps': 15512, 'loss/train': 0.5658711791038513} +01/27/2022 10:26:43 - INFO - codeparrot_training - Step 15513: {'lr': 0.0004084338350511795, 'samples': 2978688, 'steps': 15513, 'loss/train': 1.0610131323337555} +01/27/2022 10:26:46 - INFO - codeparrot_training - Step 15514: {'lr': 0.00040842117753045893, 'samples': 2978880, 'steps': 15514, 'loss/train': 0.9088510572910309} +01/27/2022 10:26:49 - INFO - codeparrot_training - Step 15515: {'lr': 0.0004084085193311124, 'samples': 2979072, 'steps': 15515, 'loss/train': 0.6154107749462128} +01/27/2022 10:26:52 - INFO - codeparrot_training - Step 15516: {'lr': 0.0004083958604531941, 'samples': 2979264, 'steps': 15516, 'loss/train': 0.6783743351697922} +01/27/2022 10:26:55 - INFO - codeparrot_training - Step 15517: {'lr': 0.0004083832008967583, 'samples': 2979456, 'steps': 15517, 'loss/train': 0.8290367424488068} +01/27/2022 10:26:59 - INFO - codeparrot_training - Step 15518: {'lr': 0.00040837054066185906, 'samples': 2979648, 'steps': 15518, 'loss/train': 0.4785505682229996} +01/27/2022 10:27:02 - INFO - codeparrot_training - Step 15519: {'lr': 0.0004083578797485508, 'samples': 2979840, 'steps': 15519, 'loss/train': 0.807129442691803} +01/27/2022 10:27:05 - INFO - codeparrot_training - Step 15520: {'lr': 0.00040834521815688753, 'samples': 2980032, 'steps': 15520, 'loss/train': 0.7868169844150543} +01/27/2022 10:27:08 - INFO - codeparrot_training - Step 15521: {'lr': 0.00040833255588692375, 'samples': 2980224, 'steps': 15521, 'loss/train': 0.6673819273710251} +01/27/2022 10:27:12 - INFO - codeparrot_training - Step 15522: {'lr': 0.0004083198929387135, 'samples': 2980416, 'steps': 15522, 'loss/train': 0.2882488891482353} +01/27/2022 10:27:16 - INFO - codeparrot_training - Step 15523: {'lr': 0.0004083072293123111, 'samples': 2980608, 'steps': 15523, 'loss/train': 0.7695490121841431} +01/27/2022 10:27:19 - INFO - codeparrot_training - Step 15524: {'lr': 0.00040829456500777084, 'samples': 2980800, 'steps': 15524, 'loss/train': 0.5395924597978592} +01/27/2022 10:27:22 - INFO - codeparrot_training - Step 15525: {'lr': 0.00040828190002514694, 'samples': 2980992, 'steps': 15525, 'loss/train': 0.39022837579250336} +01/27/2022 10:27:25 - INFO - codeparrot_training - Step 15526: {'lr': 0.0004082692343644936, 'samples': 2981184, 'steps': 15526, 'loss/train': 0.890484631061554} +01/27/2022 10:27:28 - INFO - codeparrot_training - Step 15527: {'lr': 0.00040825656802586513, 'samples': 2981376, 'steps': 15527, 'loss/train': 0.6270619332790375} +01/27/2022 10:27:31 - INFO - codeparrot_training - Step 15528: {'lr': 0.00040824390100931585, 'samples': 2981568, 'steps': 15528, 'loss/train': 0.7770337164402008} +01/27/2022 10:27:34 - INFO - codeparrot_training - Step 15529: {'lr': 0.00040823123331489985, 'samples': 2981760, 'steps': 15529, 'loss/train': 0.6430349349975586} +01/27/2022 10:27:38 - INFO - codeparrot_training - Step 15530: {'lr': 0.0004082185649426715, 'samples': 2981952, 'steps': 15530, 'loss/train': 0.26952318102121353} +01/27/2022 10:27:43 - INFO - codeparrot_training - Step 15531: {'lr': 0.0004082058958926851, 'samples': 2982144, 'steps': 15531, 'loss/train': 0.38056357204914093} +01/27/2022 10:27:46 - INFO - codeparrot_training - Step 15532: {'lr': 0.0004081932261649949, 'samples': 2982336, 'steps': 15532, 'loss/train': 0.67392897605896} +01/27/2022 10:27:49 - INFO - codeparrot_training - Step 15533: {'lr': 0.00040818055575965505, 'samples': 2982528, 'steps': 15533, 'loss/train': 1.0609095990657806} +01/27/2022 10:27:52 - INFO - codeparrot_training - Step 15534: {'lr': 0.0004081678846767199, 'samples': 2982720, 'steps': 15534, 'loss/train': 0.8146214783191681} +01/27/2022 10:27:55 - INFO - codeparrot_training - Step 15535: {'lr': 0.00040815521291624393, 'samples': 2982912, 'steps': 15535, 'loss/train': 1.0544217824935913} +01/27/2022 10:27:59 - INFO - codeparrot_training - Step 15536: {'lr': 0.0004081425404782811, 'samples': 2983104, 'steps': 15536, 'loss/train': 0.1711202748119831} +01/27/2022 10:28:02 - INFO - codeparrot_training - Step 15537: {'lr': 0.0004081298673628859, 'samples': 2983296, 'steps': 15537, 'loss/train': 0.5388596802949905} +01/27/2022 10:28:05 - INFO - codeparrot_training - Step 15538: {'lr': 0.00040811719357011257, 'samples': 2983488, 'steps': 15538, 'loss/train': 1.142158180475235} +01/27/2022 10:28:08 - INFO - codeparrot_training - Step 15539: {'lr': 0.00040810451910001537, 'samples': 2983680, 'steps': 15539, 'loss/train': 0.6383031606674194} +01/27/2022 10:28:12 - INFO - codeparrot_training - Step 15540: {'lr': 0.00040809184395264867, 'samples': 2983872, 'steps': 15540, 'loss/train': 0.9126244783401489} +01/27/2022 10:28:16 - INFO - codeparrot_training - Step 15541: {'lr': 0.0004080791681280667, 'samples': 2984064, 'steps': 15541, 'loss/train': 0.3040817230939865} +01/27/2022 10:28:19 - INFO - codeparrot_training - Step 15542: {'lr': 0.00040806649162632364, 'samples': 2984256, 'steps': 15542, 'loss/train': 1.0884831547737122} +01/27/2022 10:28:22 - INFO - codeparrot_training - Step 15543: {'lr': 0.000408053814447474, 'samples': 2984448, 'steps': 15543, 'loss/train': 0.9826526641845703} +01/27/2022 10:28:25 - INFO - codeparrot_training - Step 15544: {'lr': 0.00040804113659157203, 'samples': 2984640, 'steps': 15544, 'loss/train': 1.2122803330421448} +01/27/2022 10:28:28 - INFO - codeparrot_training - Step 15545: {'lr': 0.00040802845805867205, 'samples': 2984832, 'steps': 15545, 'loss/train': 1.0390942990779877} +01/27/2022 10:28:31 - INFO - codeparrot_training - Step 15546: {'lr': 0.0004080157788488282, 'samples': 2985024, 'steps': 15546, 'loss/train': 0.593425378203392} +01/27/2022 10:28:34 - INFO - codeparrot_training - Step 15547: {'lr': 0.0004080030989620951, 'samples': 2985216, 'steps': 15547, 'loss/train': 1.0299219489097595} +01/27/2022 10:28:38 - INFO - codeparrot_training - Step 15548: {'lr': 0.0004079904183985268, 'samples': 2985408, 'steps': 15548, 'loss/train': 0.3770473301410675} +01/27/2022 10:28:43 - INFO - codeparrot_training - Step 15549: {'lr': 0.0004079777371581777, 'samples': 2985600, 'steps': 15549, 'loss/train': 0.8531733155250549} +01/27/2022 10:28:46 - INFO - codeparrot_training - Step 15550: {'lr': 0.00040796505524110215, 'samples': 2985792, 'steps': 15550, 'loss/train': 0.507839173078537} +01/27/2022 10:28:50 - INFO - codeparrot_training - Step 15551: {'lr': 0.00040795237264735454, 'samples': 2985984, 'steps': 15551, 'loss/train': 0.5825046300888062} +01/27/2022 10:28:53 - INFO - codeparrot_training - Step 15552: {'lr': 0.00040793968937698905, 'samples': 2986176, 'steps': 15552, 'loss/train': 0.8076326251029968} +01/27/2022 10:28:56 - INFO - codeparrot_training - Step 15553: {'lr': 0.00040792700543006014, 'samples': 2986368, 'steps': 15553, 'loss/train': 1.4840116202831268} +01/27/2022 10:28:59 - INFO - codeparrot_training - Step 15554: {'lr': 0.000407914320806622, 'samples': 2986560, 'steps': 15554, 'loss/train': 0.1377776451408863} +01/27/2022 10:29:02 - INFO - codeparrot_training - Step 15555: {'lr': 0.0004079016355067291, 'samples': 2986752, 'steps': 15555, 'loss/train': 0.5085232704877853} +01/27/2022 10:29:05 - INFO - codeparrot_training - Step 15556: {'lr': 0.0004078889495304357, 'samples': 2986944, 'steps': 15556, 'loss/train': 0.3819936364889145} +01/27/2022 10:29:08 - INFO - codeparrot_training - Step 15557: {'lr': 0.00040787626287779624, 'samples': 2987136, 'steps': 15557, 'loss/train': 0.9636213183403015} +01/27/2022 10:29:13 - INFO - codeparrot_training - Step 15558: {'lr': 0.0004078635755488649, 'samples': 2987328, 'steps': 15558, 'loss/train': 0.603533998131752} +01/27/2022 10:29:16 - INFO - codeparrot_training - Step 15559: {'lr': 0.00040785088754369627, 'samples': 2987520, 'steps': 15559, 'loss/train': 0.9217486381530762} +01/27/2022 10:29:19 - INFO - codeparrot_training - Step 15560: {'lr': 0.00040783819886234445, 'samples': 2987712, 'steps': 15560, 'loss/train': 0.7269825786352158} +01/27/2022 10:29:22 - INFO - codeparrot_training - Step 15561: {'lr': 0.000407825509504864, 'samples': 2987904, 'steps': 15561, 'loss/train': 0.7753691375255585} +01/27/2022 10:29:26 - INFO - codeparrot_training - Step 15562: {'lr': 0.00040781281947130897, 'samples': 2988096, 'steps': 15562, 'loss/train': 0.8595869243144989} +01/27/2022 10:29:29 - INFO - codeparrot_training - Step 15563: {'lr': 0.0004078001287617342, 'samples': 2988288, 'steps': 15563, 'loss/train': 0.883450984954834} +01/27/2022 10:29:32 - INFO - codeparrot_training - Step 15564: {'lr': 0.0004077874373761936, 'samples': 2988480, 'steps': 15564, 'loss/train': 0.5504356473684311} +01/27/2022 10:29:35 - INFO - codeparrot_training - Step 15565: {'lr': 0.0004077747453147418, 'samples': 2988672, 'steps': 15565, 'loss/train': 0.43435701727867126} +01/27/2022 10:29:40 - INFO - codeparrot_training - Step 15566: {'lr': 0.0004077620525774331, 'samples': 2988864, 'steps': 15566, 'loss/train': 0.554634690284729} +01/27/2022 10:29:43 - INFO - codeparrot_training - Step 15567: {'lr': 0.0004077493591643219, 'samples': 2989056, 'steps': 15567, 'loss/train': 1.201582431793213} +01/27/2022 10:29:46 - INFO - codeparrot_training - Step 15568: {'lr': 0.00040773666507546244, 'samples': 2989248, 'steps': 15568, 'loss/train': 0.9120202660560608} +01/27/2022 10:29:49 - INFO - codeparrot_training - Step 15569: {'lr': 0.00040772397031090923, 'samples': 2989440, 'steps': 15569, 'loss/train': 0.6953467726707458} +01/27/2022 10:29:52 - INFO - codeparrot_training - Step 15570: {'lr': 0.0004077112748707166, 'samples': 2989632, 'steps': 15570, 'loss/train': 0.7448669672012329} +01/27/2022 10:29:55 - INFO - codeparrot_training - Step 15571: {'lr': 0.000407698578754939, 'samples': 2989824, 'steps': 15571, 'loss/train': 0.8711704015731812} +01/27/2022 10:29:58 - INFO - codeparrot_training - Step 15572: {'lr': 0.0004076858819636307, 'samples': 2990016, 'steps': 15572, 'loss/train': 0.542843297123909} +01/27/2022 10:30:02 - INFO - codeparrot_training - Step 15573: {'lr': 0.0004076731844968462, 'samples': 2990208, 'steps': 15573, 'loss/train': 0.8216398358345032} +01/27/2022 10:30:05 - INFO - codeparrot_training - Step 15574: {'lr': 0.00040766048635463984, 'samples': 2990400, 'steps': 15574, 'loss/train': 0.6938819736242294} +01/27/2022 10:30:11 - INFO - codeparrot_training - Step 15575: {'lr': 0.000407647787537066, 'samples': 2990592, 'steps': 15575, 'loss/train': 5.0880206823349} +01/27/2022 10:30:14 - INFO - codeparrot_training - Step 15576: {'lr': 0.00040763508804417904, 'samples': 2990784, 'steps': 15576, 'loss/train': 5.035927176475525} +01/27/2022 10:30:17 - INFO - codeparrot_training - Step 15577: {'lr': 0.0004076223878760335, 'samples': 2990976, 'steps': 15577, 'loss/train': 0.8472124636173248} +01/27/2022 10:30:20 - INFO - codeparrot_training - Step 15578: {'lr': 0.0004076096870326837, 'samples': 2991168, 'steps': 15578, 'loss/train': 0.4127427041530609} +01/27/2022 10:30:23 - INFO - codeparrot_training - Step 15579: {'lr': 0.000407596985514184, 'samples': 2991360, 'steps': 15579, 'loss/train': 0.8911983668804169} +01/27/2022 10:30:26 - INFO - codeparrot_training - Step 15580: {'lr': 0.00040758428332058895, 'samples': 2991552, 'steps': 15580, 'loss/train': 0.4939490854740143} +01/27/2022 10:30:29 - INFO - codeparrot_training - Step 15581: {'lr': 0.00040757158045195274, 'samples': 2991744, 'steps': 15581, 'loss/train': 1.0463954508304596} +01/27/2022 10:30:33 - INFO - codeparrot_training - Step 15582: {'lr': 0.00040755887690833005, 'samples': 2991936, 'steps': 15582, 'loss/train': 1.027804970741272} +01/27/2022 10:30:36 - INFO - codeparrot_training - Step 15583: {'lr': 0.00040754617268977503, 'samples': 2992128, 'steps': 15583, 'loss/train': 0.5191497355699539} +01/27/2022 10:30:40 - INFO - codeparrot_training - Step 15584: {'lr': 0.0004075334677963423, 'samples': 2992320, 'steps': 15584, 'loss/train': 0.8034518659114838} +01/27/2022 10:30:43 - INFO - codeparrot_training - Step 15585: {'lr': 0.00040752076222808623, 'samples': 2992512, 'steps': 15585, 'loss/train': 0.12912015616893768} +01/27/2022 10:30:47 - INFO - codeparrot_training - Step 15586: {'lr': 0.00040750805598506115, 'samples': 2992704, 'steps': 15586, 'loss/train': 0.7208998650312424} +01/27/2022 10:30:50 - INFO - codeparrot_training - Step 15587: {'lr': 0.00040749534906732167, 'samples': 2992896, 'steps': 15587, 'loss/train': 0.5922729671001434} +01/27/2022 10:30:53 - INFO - codeparrot_training - Step 15588: {'lr': 0.0004074826414749221, 'samples': 2993088, 'steps': 15588, 'loss/train': 0.5022197216749191} +01/27/2022 10:30:56 - INFO - codeparrot_training - Step 15589: {'lr': 0.00040746993320791685, 'samples': 2993280, 'steps': 15589, 'loss/train': 0.38245511054992676} +01/27/2022 10:30:59 - INFO - codeparrot_training - Step 15590: {'lr': 0.00040745722426636043, 'samples': 2993472, 'steps': 15590, 'loss/train': 0.8424258828163147} +01/27/2022 10:31:02 - INFO - codeparrot_training - Step 15591: {'lr': 0.0004074445146503073, 'samples': 2993664, 'steps': 15591, 'loss/train': 0.6391682177782059} +01/27/2022 10:31:06 - INFO - codeparrot_training - Step 15592: {'lr': 0.00040743180435981187, 'samples': 2993856, 'steps': 15592, 'loss/train': 0.6918118894100189} +01/27/2022 10:31:10 - INFO - codeparrot_training - Step 15593: {'lr': 0.0004074190933949286, 'samples': 2994048, 'steps': 15593, 'loss/train': 0.5519225299358368} +01/27/2022 10:31:13 - INFO - codeparrot_training - Step 15594: {'lr': 0.00040740638175571175, 'samples': 2994240, 'steps': 15594, 'loss/train': 0.387773260474205} +01/27/2022 10:31:17 - INFO - codeparrot_training - Step 15595: {'lr': 0.0004073936694422161, 'samples': 2994432, 'steps': 15595, 'loss/train': 0.789720743894577} +01/27/2022 10:31:20 - INFO - codeparrot_training - Step 15596: {'lr': 0.0004073809564544959, 'samples': 2994624, 'steps': 15596, 'loss/train': 0.06723910197615623} +01/27/2022 10:31:23 - INFO - codeparrot_training - Step 15597: {'lr': 0.0004073682427926057, 'samples': 2994816, 'steps': 15597, 'loss/train': 0.9839837551116943} +01/27/2022 10:31:26 - INFO - codeparrot_training - Step 15598: {'lr': 0.00040735552845659986, 'samples': 2995008, 'steps': 15598, 'loss/train': 0.8038419485092163} +01/27/2022 10:31:29 - INFO - codeparrot_training - Step 15599: {'lr': 0.00040734281344653294, 'samples': 2995200, 'steps': 15599, 'loss/train': 0.9281355142593384} +01/27/2022 10:31:32 - INFO - codeparrot_training - Step 15600: {'lr': 0.0004073300977624594, 'samples': 2995392, 'steps': 15600, 'loss/train': 0.5337085872888565} +01/27/2022 10:31:36 - INFO - codeparrot_training - Step 15601: {'lr': 0.0004073173814044336, 'samples': 2995584, 'steps': 15601, 'loss/train': 0.7302671670913696} +01/27/2022 10:31:40 - INFO - codeparrot_training - Step 15602: {'lr': 0.0004073046643725101, 'samples': 2995776, 'steps': 15602, 'loss/train': 0.8859627842903137} +01/27/2022 10:31:43 - INFO - codeparrot_training - Step 15603: {'lr': 0.0004072919466667434, 'samples': 2995968, 'steps': 15603, 'loss/train': 1.456421971321106} +01/27/2022 10:31:46 - INFO - codeparrot_training - Step 15604: {'lr': 0.000407279228287188, 'samples': 2996160, 'steps': 15604, 'loss/train': 0.6069286018610001} +01/27/2022 10:31:49 - INFO - codeparrot_training - Step 15605: {'lr': 0.00040726650923389825, 'samples': 2996352, 'steps': 15605, 'loss/train': 0.8599977493286133} +01/27/2022 10:31:53 - INFO - codeparrot_training - Step 15606: {'lr': 0.00040725378950692874, 'samples': 2996544, 'steps': 15606, 'loss/train': 0.6981897801160812} +01/27/2022 10:31:56 - INFO - codeparrot_training - Step 15607: {'lr': 0.0004072410691063339, 'samples': 2996736, 'steps': 15607, 'loss/train': 0.973052054643631} +01/27/2022 10:31:59 - INFO - codeparrot_training - Step 15608: {'lr': 0.00040722834803216834, 'samples': 2996928, 'steps': 15608, 'loss/train': 0.6552781462669373} +01/27/2022 10:32:02 - INFO - codeparrot_training - Step 15609: {'lr': 0.0004072156262844864, 'samples': 2997120, 'steps': 15609, 'loss/train': 1.1246516704559326} +01/27/2022 10:32:07 - INFO - codeparrot_training - Step 15610: {'lr': 0.0004072029038633426, 'samples': 2997312, 'steps': 15610, 'loss/train': 0.4990398734807968} +01/27/2022 10:32:10 - INFO - codeparrot_training - Step 15611: {'lr': 0.0004071901807687915, 'samples': 2997504, 'steps': 15611, 'loss/train': 0.7242873162031174} +01/27/2022 10:32:13 - INFO - codeparrot_training - Step 15612: {'lr': 0.0004071774570008876, 'samples': 2997696, 'steps': 15612, 'loss/train': 0.6396433711051941} +01/27/2022 10:32:17 - INFO - codeparrot_training - Step 15613: {'lr': 0.00040716473255968534, 'samples': 2997888, 'steps': 15613, 'loss/train': 0.6063070446252823} +01/27/2022 10:32:20 - INFO - codeparrot_training - Step 15614: {'lr': 0.0004071520074452393, 'samples': 2998080, 'steps': 15614, 'loss/train': 1.178872436285019} +01/27/2022 10:32:23 - INFO - codeparrot_training - Step 15615: {'lr': 0.000407139281657604, 'samples': 2998272, 'steps': 15615, 'loss/train': 0.3146365284919739} +01/27/2022 10:32:26 - INFO - codeparrot_training - Step 15616: {'lr': 0.0004071265551968338, 'samples': 2998464, 'steps': 15616, 'loss/train': 0.6595892161130905} +01/27/2022 10:32:29 - INFO - codeparrot_training - Step 15617: {'lr': 0.0004071138280629835, 'samples': 2998656, 'steps': 15617, 'loss/train': 1.1717610955238342} +01/27/2022 10:32:32 - INFO - codeparrot_training - Step 15618: {'lr': 0.00040710110025610733, 'samples': 2998848, 'steps': 15618, 'loss/train': 0.7536022961139679} +01/27/2022 10:32:37 - INFO - codeparrot_training - Step 15619: {'lr': 0.00040708837177626, 'samples': 2999040, 'steps': 15619, 'loss/train': 0.9073634147644043} +01/27/2022 10:32:40 - INFO - codeparrot_training - Step 15620: {'lr': 0.00040707564262349594, 'samples': 2999232, 'steps': 15620, 'loss/train': 0.28159095346927643} +01/27/2022 10:32:43 - INFO - codeparrot_training - Step 15621: {'lr': 0.00040706291279786965, 'samples': 2999424, 'steps': 15621, 'loss/train': 0.28899912536144257} +01/27/2022 10:32:46 - INFO - codeparrot_training - Step 15622: {'lr': 0.0004070501822994358, 'samples': 2999616, 'steps': 15622, 'loss/train': 0.5352132618427277} +01/27/2022 10:32:50 - INFO - codeparrot_training - Step 15623: {'lr': 0.00040703745112824876, 'samples': 2999808, 'steps': 15623, 'loss/train': 0.3112994357943535} +01/27/2022 10:32:53 - INFO - codeparrot_training - Step 15624: {'lr': 0.00040702471928436316, 'samples': 3000000, 'steps': 15624, 'loss/train': 0.8546259999275208} +01/27/2022 10:32:56 - INFO - codeparrot_training - Step 15625: {'lr': 0.00040701198676783355, 'samples': 3000192, 'steps': 15625, 'loss/train': 0.14668841287493706} +01/27/2022 10:32:59 - INFO - codeparrot_training - Step 15626: {'lr': 0.00040699925357871446, 'samples': 3000384, 'steps': 15626, 'loss/train': 1.291738897562027} +01/27/2022 10:33:02 - INFO - codeparrot_training - Step 15627: {'lr': 0.00040698651971706037, 'samples': 3000576, 'steps': 15627, 'loss/train': 0.4575551301240921} +01/27/2022 10:33:07 - INFO - codeparrot_training - Step 15628: {'lr': 0.00040697378518292593, 'samples': 3000768, 'steps': 15628, 'loss/train': 0.9747738540172577} +01/27/2022 10:33:10 - INFO - codeparrot_training - Step 15629: {'lr': 0.0004069610499763656, 'samples': 3000960, 'steps': 15629, 'loss/train': 0.7424184679985046} +01/27/2022 10:33:13 - INFO - codeparrot_training - Step 15630: {'lr': 0.00040694831409743406, 'samples': 3001152, 'steps': 15630, 'loss/train': 0.7376485168933868} +01/27/2022 10:33:16 - INFO - codeparrot_training - Step 15631: {'lr': 0.00040693557754618566, 'samples': 3001344, 'steps': 15631, 'loss/train': 0.4232730567455292} +01/27/2022 10:33:19 - INFO - codeparrot_training - Step 15632: {'lr': 0.00040692284032267515, 'samples': 3001536, 'steps': 15632, 'loss/train': 0.03516738396137953} +01/27/2022 10:33:23 - INFO - codeparrot_training - Step 15633: {'lr': 0.00040691010242695696, 'samples': 3001728, 'steps': 15633, 'loss/train': 0.78569296002388} +01/27/2022 10:33:26 - INFO - codeparrot_training - Step 15634: {'lr': 0.00040689736385908574, 'samples': 3001920, 'steps': 15634, 'loss/train': 0.7876304984092712} +01/27/2022 10:33:29 - INFO - codeparrot_training - Step 15635: {'lr': 0.0004068846246191161, 'samples': 3002112, 'steps': 15635, 'loss/train': 0.5231411755084991} +01/27/2022 10:33:32 - INFO - codeparrot_training - Step 15636: {'lr': 0.00040687188470710245, 'samples': 3002304, 'steps': 15636, 'loss/train': 1.5908406972885132} +01/27/2022 10:33:38 - INFO - codeparrot_training - Step 15637: {'lr': 0.00040685914412309955, 'samples': 3002496, 'steps': 15637, 'loss/train': 0.6664626449346542} +01/27/2022 10:33:41 - INFO - codeparrot_training - Step 15638: {'lr': 0.0004068464028671618, 'samples': 3002688, 'steps': 15638, 'loss/train': 0.5473641157150269} +01/27/2022 10:33:44 - INFO - codeparrot_training - Step 15639: {'lr': 0.00040683366093934394, 'samples': 3002880, 'steps': 15639, 'loss/train': 0.45102958381175995} +01/27/2022 10:33:47 - INFO - codeparrot_training - Step 15640: {'lr': 0.0004068209183397004, 'samples': 3003072, 'steps': 15640, 'loss/train': 0.9850993752479553} +01/27/2022 10:33:50 - INFO - codeparrot_training - Step 15641: {'lr': 0.0004068081750682859, 'samples': 3003264, 'steps': 15641, 'loss/train': 0.8080179691314697} +01/27/2022 10:33:53 - INFO - codeparrot_training - Step 15642: {'lr': 0.00040679543112515494, 'samples': 3003456, 'steps': 15642, 'loss/train': 0.7773769497871399} +01/27/2022 10:33:56 - INFO - codeparrot_training - Step 15643: {'lr': 0.00040678268651036213, 'samples': 3003648, 'steps': 15643, 'loss/train': 0.8158512711524963} +01/27/2022 10:34:00 - INFO - codeparrot_training - Step 15644: {'lr': 0.0004067699412239622, 'samples': 3003840, 'steps': 15644, 'loss/train': 0.7840842604637146} +01/27/2022 10:34:03 - INFO - codeparrot_training - Step 15645: {'lr': 0.00040675719526600947, 'samples': 3004032, 'steps': 15645, 'loss/train': 0.036787248216569424} +01/27/2022 10:34:07 - INFO - codeparrot_training - Step 15646: {'lr': 0.0004067444486365587, 'samples': 3004224, 'steps': 15646, 'loss/train': 1.0905160009860992} +01/27/2022 10:34:10 - INFO - codeparrot_training - Step 15647: {'lr': 0.00040673170133566453, 'samples': 3004416, 'steps': 15647, 'loss/train': 0.5925396233797073} +01/27/2022 10:34:14 - INFO - codeparrot_training - Step 15648: {'lr': 0.0004067189533633815, 'samples': 3004608, 'steps': 15648, 'loss/train': 1.0180284976959229} +01/27/2022 10:34:17 - INFO - codeparrot_training - Step 15649: {'lr': 0.00040670620471976426, 'samples': 3004800, 'steps': 15649, 'loss/train': 1.4313391149044037} +01/27/2022 10:34:20 - INFO - codeparrot_training - Step 15650: {'lr': 0.0004066934554048674, 'samples': 3004992, 'steps': 15650, 'loss/train': 0.9943420886993408} +01/27/2022 10:34:23 - INFO - codeparrot_training - Step 15651: {'lr': 0.00040668070541874553, 'samples': 3005184, 'steps': 15651, 'loss/train': 0.6132618337869644} +01/27/2022 10:34:26 - INFO - codeparrot_training - Step 15652: {'lr': 0.00040666795476145326, 'samples': 3005376, 'steps': 15652, 'loss/train': 1.041513741016388} +01/27/2022 10:34:29 - INFO - codeparrot_training - Step 15653: {'lr': 0.00040665520343304516, 'samples': 3005568, 'steps': 15653, 'loss/train': 0.6273356527090073} +01/27/2022 10:34:34 - INFO - codeparrot_training - Step 15654: {'lr': 0.00040664245143357604, 'samples': 3005760, 'steps': 15654, 'loss/train': 1.182309776544571} +01/27/2022 10:34:38 - INFO - codeparrot_training - Step 15655: {'lr': 0.0004066296987631003, 'samples': 3005952, 'steps': 15655, 'loss/train': 0.8292187750339508} +01/27/2022 10:34:41 - INFO - codeparrot_training - Step 15656: {'lr': 0.0004066169454216727, 'samples': 3006144, 'steps': 15656, 'loss/train': 0.621228888630867} +01/27/2022 10:34:44 - INFO - codeparrot_training - Step 15657: {'lr': 0.00040660419140934787, 'samples': 3006336, 'steps': 15657, 'loss/train': 1.289659470319748} +01/27/2022 10:34:47 - INFO - codeparrot_training - Step 15658: {'lr': 0.0004065914367261804, 'samples': 3006528, 'steps': 15658, 'loss/train': 0.7034256756305695} +01/27/2022 10:34:50 - INFO - codeparrot_training - Step 15659: {'lr': 0.00040657868137222486, 'samples': 3006720, 'steps': 15659, 'loss/train': 0.9773639738559723} +01/27/2022 10:34:53 - INFO - codeparrot_training - Step 15660: {'lr': 0.000406565925347536, 'samples': 3006912, 'steps': 15660, 'loss/train': 0.3563143461942673} +01/27/2022 10:34:57 - INFO - codeparrot_training - Step 15661: {'lr': 0.0004065531686521685, 'samples': 3007104, 'steps': 15661, 'loss/train': 0.7162356823682785} +01/27/2022 10:35:00 - INFO - codeparrot_training - Step 15662: {'lr': 0.00040654041128617693, 'samples': 3007296, 'steps': 15662, 'loss/train': 0.7455568313598633} +01/27/2022 10:35:04 - INFO - codeparrot_training - Step 15663: {'lr': 0.0004065276532496158, 'samples': 3007488, 'steps': 15663, 'loss/train': 0.6190991252660751} +01/27/2022 10:35:07 - INFO - codeparrot_training - Step 15664: {'lr': 0.0004065148945425401, 'samples': 3007680, 'steps': 15664, 'loss/train': 0.6893211901187897} +01/27/2022 10:35:10 - INFO - codeparrot_training - Step 15665: {'lr': 0.0004065021351650042, 'samples': 3007872, 'steps': 15665, 'loss/train': 1.2268229126930237} +01/27/2022 10:35:14 - INFO - codeparrot_training - Step 15666: {'lr': 0.00040648937511706285, 'samples': 3008064, 'steps': 15666, 'loss/train': 1.2060426771640778} +01/27/2022 10:35:17 - INFO - codeparrot_training - Step 15667: {'lr': 0.0004064766143987707, 'samples': 3008256, 'steps': 15667, 'loss/train': 0.6545256525278091} +01/27/2022 10:35:20 - INFO - codeparrot_training - Step 15668: {'lr': 0.00040646385301018243, 'samples': 3008448, 'steps': 15668, 'loss/train': 1.3056898713111877} +01/27/2022 10:35:23 - INFO - codeparrot_training - Step 15669: {'lr': 0.0004064510909513527, 'samples': 3008640, 'steps': 15669, 'loss/train': 0.7700411081314087} +01/27/2022 10:35:26 - INFO - codeparrot_training - Step 15670: {'lr': 0.00040643832822233615, 'samples': 3008832, 'steps': 15670, 'loss/train': 0.7135010361671448} +01/27/2022 10:35:29 - INFO - codeparrot_training - Step 15671: {'lr': 0.0004064255648231875, 'samples': 3009024, 'steps': 15671, 'loss/train': 0.49279242753982544} +01/27/2022 10:35:34 - INFO - codeparrot_training - Step 15672: {'lr': 0.00040641280075396144, 'samples': 3009216, 'steps': 15672, 'loss/train': 1.022241622209549} +01/27/2022 10:35:37 - INFO - codeparrot_training - Step 15673: {'lr': 0.00040640003601471255, 'samples': 3009408, 'steps': 15673, 'loss/train': 0.8535444438457489} +01/27/2022 10:35:40 - INFO - codeparrot_training - Step 15674: {'lr': 0.00040638727060549556, 'samples': 3009600, 'steps': 15674, 'loss/train': 0.6760061681270599} +01/27/2022 10:35:43 - INFO - codeparrot_training - Step 15675: {'lr': 0.00040637450452636517, 'samples': 3009792, 'steps': 15675, 'loss/train': 0.7199352979660034} +01/27/2022 10:35:46 - INFO - codeparrot_training - Step 15676: {'lr': 0.00040636173777737613, 'samples': 3009984, 'steps': 15676, 'loss/train': 0.8876660764217377} +01/27/2022 10:35:50 - INFO - codeparrot_training - Step 15677: {'lr': 0.000406348970358583, 'samples': 3010176, 'steps': 15677, 'loss/train': 0.48838184773921967} +01/27/2022 10:35:53 - INFO - codeparrot_training - Step 15678: {'lr': 0.00040633620227004054, 'samples': 3010368, 'steps': 15678, 'loss/train': 0.7037598788738251} +01/27/2022 10:35:56 - INFO - codeparrot_training - Step 15679: {'lr': 0.0004063234335118033, 'samples': 3010560, 'steps': 15679, 'loss/train': 0.7762987911701202} +01/27/2022 10:35:59 - INFO - codeparrot_training - Step 15680: {'lr': 0.00040631066408392636, 'samples': 3010752, 'steps': 15680, 'loss/train': 0.5776901543140411} +01/27/2022 10:36:05 - INFO - codeparrot_training - Step 15681: {'lr': 0.000406297893986464, 'samples': 3010944, 'steps': 15681, 'loss/train': 0.5503857135772705} +01/27/2022 10:36:08 - INFO - codeparrot_training - Step 15682: {'lr': 0.0004062851232194711, 'samples': 3011136, 'steps': 15682, 'loss/train': 0.45200349390506744} +01/27/2022 10:36:11 - INFO - codeparrot_training - Step 15683: {'lr': 0.00040627235178300236, 'samples': 3011328, 'steps': 15683, 'loss/train': 0.5987409353256226} +01/27/2022 10:36:14 - INFO - codeparrot_training - Step 15684: {'lr': 0.0004062595796771126, 'samples': 3011520, 'steps': 15684, 'loss/train': 0.5401462465524673} +01/27/2022 10:36:17 - INFO - codeparrot_training - Step 15685: {'lr': 0.0004062468069018563, 'samples': 3011712, 'steps': 15685, 'loss/train': 0.9766831398010254} +01/27/2022 10:36:21 - INFO - codeparrot_training - Step 15686: {'lr': 0.0004062340334572883, 'samples': 3011904, 'steps': 15686, 'loss/train': 0.020332693587988615} +01/27/2022 10:36:24 - INFO - codeparrot_training - Step 15687: {'lr': 0.0004062212593434634, 'samples': 3012096, 'steps': 15687, 'loss/train': 0.2760489508509636} +01/27/2022 10:36:27 - INFO - codeparrot_training - Step 15688: {'lr': 0.0004062084845604361, 'samples': 3012288, 'steps': 15688, 'loss/train': 0.10797799006104469} +01/27/2022 10:36:32 - INFO - codeparrot_training - Step 15689: {'lr': 0.00040619570910826135, 'samples': 3012480, 'steps': 15689, 'loss/train': 0.43655043840408325} +01/27/2022 10:36:35 - INFO - codeparrot_training - Step 15690: {'lr': 0.0004061829329869937, 'samples': 3012672, 'steps': 15690, 'loss/train': 1.0677664875984192} +01/27/2022 10:36:38 - INFO - codeparrot_training - Step 15691: {'lr': 0.0004061701561966881, 'samples': 3012864, 'steps': 15691, 'loss/train': 0.8104136288166046} +01/27/2022 10:36:41 - INFO - codeparrot_training - Step 15692: {'lr': 0.000406157378737399, 'samples': 3013056, 'steps': 15692, 'loss/train': 1.7483943700790405} +01/27/2022 10:36:44 - INFO - codeparrot_training - Step 15693: {'lr': 0.00040614460060918136, 'samples': 3013248, 'steps': 15693, 'loss/train': 0.8786450028419495} +01/27/2022 10:36:47 - INFO - codeparrot_training - Step 15694: {'lr': 0.0004061318218120898, 'samples': 3013440, 'steps': 15694, 'loss/train': 0.9270368814468384} +01/27/2022 10:36:50 - INFO - codeparrot_training - Step 15695: {'lr': 0.000406119042346179, 'samples': 3013632, 'steps': 15695, 'loss/train': 0.8598174154758453} +01/27/2022 10:36:54 - INFO - codeparrot_training - Step 15696: {'lr': 0.0004061062622115039, 'samples': 3013824, 'steps': 15696, 'loss/train': 0.9367427229881287} +01/27/2022 10:36:57 - INFO - codeparrot_training - Step 15697: {'lr': 0.0004060934814081192, 'samples': 3014016, 'steps': 15697, 'loss/train': 1.5508900880813599} +01/27/2022 10:37:00 - INFO - codeparrot_training - Step 15698: {'lr': 0.00040608069993607954, 'samples': 3014208, 'steps': 15698, 'loss/train': 1.8156405687332153} +01/27/2022 10:37:04 - INFO - codeparrot_training - Step 15699: {'lr': 0.00040606791779543966, 'samples': 3014400, 'steps': 15699, 'loss/train': 1.9201408624649048} +01/27/2022 10:37:07 - INFO - codeparrot_training - Step 15700: {'lr': 0.00040605513498625443, 'samples': 3014592, 'steps': 15700, 'loss/train': 1.0708368122577667} +01/27/2022 10:37:11 - INFO - codeparrot_training - Step 15701: {'lr': 0.00040604235150857855, 'samples': 3014784, 'steps': 15701, 'loss/train': 1.0034924447536469} +01/27/2022 10:37:14 - INFO - codeparrot_training - Step 15702: {'lr': 0.00040602956736246677, 'samples': 3014976, 'steps': 15702, 'loss/train': 1.296532541513443} +01/27/2022 10:37:17 - INFO - codeparrot_training - Step 15703: {'lr': 0.00040601678254797394, 'samples': 3015168, 'steps': 15703, 'loss/train': 0.604518860578537} +01/27/2022 10:37:20 - INFO - codeparrot_training - Step 15704: {'lr': 0.00040600399706515466, 'samples': 3015360, 'steps': 15704, 'loss/train': 1.0621517300605774} +01/27/2022 10:37:23 - INFO - codeparrot_training - Step 15705: {'lr': 0.0004059912109140638, 'samples': 3015552, 'steps': 15705, 'loss/train': 0.3516403064131737} +01/27/2022 10:37:26 - INFO - codeparrot_training - Step 15706: {'lr': 0.00040597842409475615, 'samples': 3015744, 'steps': 15706, 'loss/train': 0.5540933310985565} +01/27/2022 10:37:30 - INFO - codeparrot_training - Step 15707: {'lr': 0.00040596563660728646, 'samples': 3015936, 'steps': 15707, 'loss/train': 0.9122962653636932} +01/27/2022 10:37:34 - INFO - codeparrot_training - Step 15708: {'lr': 0.00040595284845170956, 'samples': 3016128, 'steps': 15708, 'loss/train': 0.9515817761421204} +01/27/2022 10:37:38 - INFO - codeparrot_training - Step 15709: {'lr': 0.0004059400596280801, 'samples': 3016320, 'steps': 15709, 'loss/train': 0.9576432108879089} +01/27/2022 10:37:41 - INFO - codeparrot_training - Step 15710: {'lr': 0.00040592727013645297, 'samples': 3016512, 'steps': 15710, 'loss/train': 0.5693067312240601} +01/27/2022 10:37:44 - INFO - codeparrot_training - Step 15711: {'lr': 0.0004059144799768829, 'samples': 3016704, 'steps': 15711, 'loss/train': 0.9694055914878845} +01/27/2022 10:37:47 - INFO - codeparrot_training - Step 15712: {'lr': 0.00040590168914942477, 'samples': 3016896, 'steps': 15712, 'loss/train': 0.7975320518016815} +01/27/2022 10:37:50 - INFO - codeparrot_training - Step 15713: {'lr': 0.0004058888976541333, 'samples': 3017088, 'steps': 15713, 'loss/train': 1.1337738633155823} +01/27/2022 10:37:53 - INFO - codeparrot_training - Step 15714: {'lr': 0.00040587610549106326, 'samples': 3017280, 'steps': 15714, 'loss/train': 0.8402758240699768} +01/27/2022 10:37:56 - INFO - codeparrot_training - Step 15715: {'lr': 0.00040586331266026943, 'samples': 3017472, 'steps': 15715, 'loss/train': 0.16281845793128014} +01/27/2022 10:37:59 - INFO - codeparrot_training - Step 15716: {'lr': 0.0004058505191618067, 'samples': 3017664, 'steps': 15716, 'loss/train': 0.8250508904457092} +01/27/2022 10:38:05 - INFO - codeparrot_training - Step 15717: {'lr': 0.0004058377249957299, 'samples': 3017856, 'steps': 15717, 'loss/train': 1.0698883831501007} +01/27/2022 10:38:08 - INFO - codeparrot_training - Step 15718: {'lr': 0.0004058249301620937, 'samples': 3018048, 'steps': 15718, 'loss/train': 0.7564921975135803} +01/27/2022 10:38:11 - INFO - codeparrot_training - Step 15719: {'lr': 0.00040581213466095304, 'samples': 3018240, 'steps': 15719, 'loss/train': 0.8794682621955872} +01/27/2022 10:38:14 - INFO - codeparrot_training - Step 15720: {'lr': 0.0004057993384923626, 'samples': 3018432, 'steps': 15720, 'loss/train': 0.6597152799367905} +01/27/2022 10:38:17 - INFO - codeparrot_training - Step 15721: {'lr': 0.0004057865416563773, 'samples': 3018624, 'steps': 15721, 'loss/train': 1.000141203403473} +01/27/2022 10:38:20 - INFO - codeparrot_training - Step 15722: {'lr': 0.0004057737441530519, 'samples': 3018816, 'steps': 15722, 'loss/train': 0.8694716691970825} +01/27/2022 10:38:24 - INFO - codeparrot_training - Step 15723: {'lr': 0.0004057609459824412, 'samples': 3019008, 'steps': 15723, 'loss/train': 0.6742167323827744} +01/27/2022 10:38:27 - INFO - codeparrot_training - Step 15724: {'lr': 0.00040574814714460015, 'samples': 3019200, 'steps': 15724, 'loss/train': 0.8840144276618958} +01/27/2022 10:38:31 - INFO - codeparrot_training - Step 15725: {'lr': 0.0004057353476395835, 'samples': 3019392, 'steps': 15725, 'loss/train': 0.870579332113266} +01/27/2022 10:38:35 - INFO - codeparrot_training - Step 15726: {'lr': 0.00040572254746744607, 'samples': 3019584, 'steps': 15726, 'loss/train': 0.6343060433864594} +01/27/2022 10:38:38 - INFO - codeparrot_training - Step 15727: {'lr': 0.00040570974662824266, 'samples': 3019776, 'steps': 15727, 'loss/train': 0.5992527455091476} +01/27/2022 10:38:41 - INFO - codeparrot_training - Step 15728: {'lr': 0.00040569694512202815, 'samples': 3019968, 'steps': 15728, 'loss/train': 0.689862459897995} +01/27/2022 10:38:44 - INFO - codeparrot_training - Step 15729: {'lr': 0.00040568414294885736, 'samples': 3020160, 'steps': 15729, 'loss/train': 0.3691251575946808} +01/27/2022 10:38:47 - INFO - codeparrot_training - Step 15730: {'lr': 0.00040567134010878513, 'samples': 3020352, 'steps': 15730, 'loss/train': 0.6531042605638504} +01/27/2022 10:38:50 - INFO - codeparrot_training - Step 15731: {'lr': 0.00040565853660186633, 'samples': 3020544, 'steps': 15731, 'loss/train': 0.9442330598831177} +01/27/2022 10:38:53 - INFO - codeparrot_training - Step 15732: {'lr': 0.0004056457324281557, 'samples': 3020736, 'steps': 15732, 'loss/train': 1.929390549659729} +01/27/2022 10:38:57 - INFO - codeparrot_training - Step 15733: {'lr': 0.0004056329275877083, 'samples': 3020928, 'steps': 15733, 'loss/train': 1.1133512556552887} +01/27/2022 10:39:02 - INFO - codeparrot_training - Step 15734: {'lr': 0.00040562012208057886, 'samples': 3021120, 'steps': 15734, 'loss/train': 0.7730762958526611} +01/27/2022 10:39:05 - INFO - codeparrot_training - Step 15735: {'lr': 0.0004056073159068222, 'samples': 3021312, 'steps': 15735, 'loss/train': 1.1235328316688538} +01/27/2022 10:39:08 - INFO - codeparrot_training - Step 15736: {'lr': 0.0004055945090664931, 'samples': 3021504, 'steps': 15736, 'loss/train': 0.594909131526947} +01/27/2022 10:39:11 - INFO - codeparrot_training - Step 15737: {'lr': 0.0004055817015596467, 'samples': 3021696, 'steps': 15737, 'loss/train': 0.6950836926698685} +01/27/2022 10:39:14 - INFO - codeparrot_training - Step 15738: {'lr': 0.00040556889338633754, 'samples': 3021888, 'steps': 15738, 'loss/train': 0.5858600735664368} +01/27/2022 10:39:17 - INFO - codeparrot_training - Step 15739: {'lr': 0.00040555608454662074, 'samples': 3022080, 'steps': 15739, 'loss/train': 1.0516335368156433} +01/27/2022 10:39:21 - INFO - codeparrot_training - Step 15740: {'lr': 0.00040554327504055106, 'samples': 3022272, 'steps': 15740, 'loss/train': 1.1074225008487701} +01/27/2022 10:39:24 - INFO - codeparrot_training - Step 15741: {'lr': 0.00040553046486818336, 'samples': 3022464, 'steps': 15741, 'loss/train': 0.6780335158109665} +01/27/2022 10:39:27 - INFO - codeparrot_training - Step 15742: {'lr': 0.0004055176540295725, 'samples': 3022656, 'steps': 15742, 'loss/train': 1.765175700187683} +01/27/2022 10:39:31 - INFO - codeparrot_training - Step 15743: {'lr': 0.00040550484252477347, 'samples': 3022848, 'steps': 15743, 'loss/train': 0.8886852264404297} +01/27/2022 10:39:34 - INFO - codeparrot_training - Step 15744: {'lr': 0.00040549203035384105, 'samples': 3023040, 'steps': 15744, 'loss/train': 0.5455527752637863} +01/27/2022 10:39:38 - INFO - codeparrot_training - Step 15745: {'lr': 0.0004054792175168301, 'samples': 3023232, 'steps': 15745, 'loss/train': 0.9437538385391235} +01/27/2022 10:39:41 - INFO - codeparrot_training - Step 15746: {'lr': 0.00040546640401379556, 'samples': 3023424, 'steps': 15746, 'loss/train': 0.8903069794178009} +01/27/2022 10:39:44 - INFO - codeparrot_training - Step 15747: {'lr': 0.0004054535898447924, 'samples': 3023616, 'steps': 15747, 'loss/train': 0.19941667467355728} +01/27/2022 10:39:47 - INFO - codeparrot_training - Step 15748: {'lr': 0.0004054407750098753, 'samples': 3023808, 'steps': 15748, 'loss/train': 0.7813836336135864} +01/27/2022 10:39:50 - INFO - codeparrot_training - Step 15749: {'lr': 0.0004054279595090994, 'samples': 3024000, 'steps': 15749, 'loss/train': 0.9351506531238556} +01/27/2022 10:39:53 - INFO - codeparrot_training - Step 15750: {'lr': 0.0004054151433425194, 'samples': 3024192, 'steps': 15750, 'loss/train': 1.1451847851276398} +01/27/2022 10:39:56 - INFO - codeparrot_training - Step 15751: {'lr': 0.00040540232651019027, 'samples': 3024384, 'steps': 15751, 'loss/train': 1.0219130516052246} +01/27/2022 10:40:01 - INFO - codeparrot_training - Step 15752: {'lr': 0.0004053895090121669, 'samples': 3024576, 'steps': 15752, 'loss/train': 1.0680803954601288} +01/27/2022 10:40:04 - INFO - codeparrot_training - Step 15753: {'lr': 0.00040537669084850426, 'samples': 3024768, 'steps': 15753, 'loss/train': 1.1061465740203857} +01/27/2022 10:40:07 - INFO - codeparrot_training - Step 15754: {'lr': 0.0004053638720192572, 'samples': 3024960, 'steps': 15754, 'loss/train': 0.2594103142619133} +01/27/2022 10:40:10 - INFO - codeparrot_training - Step 15755: {'lr': 0.00040535105252448067, 'samples': 3025152, 'steps': 15755, 'loss/train': 1.0529007017612457} +01/27/2022 10:40:13 - INFO - codeparrot_training - Step 15756: {'lr': 0.0004053382323642295, 'samples': 3025344, 'steps': 15756, 'loss/train': 0.9301646947860718} +01/27/2022 10:40:17 - INFO - codeparrot_training - Step 15757: {'lr': 0.0004053254115385587, 'samples': 3025536, 'steps': 15757, 'loss/train': 0.6411663293838501} +01/27/2022 10:40:20 - INFO - codeparrot_training - Step 15758: {'lr': 0.00040531259004752317, 'samples': 3025728, 'steps': 15758, 'loss/train': 0.19054855406284332} +01/27/2022 10:40:23 - INFO - codeparrot_training - Step 15759: {'lr': 0.00040529976789117786, 'samples': 3025920, 'steps': 15759, 'loss/train': 0.8676028847694397} +01/27/2022 10:40:26 - INFO - codeparrot_training - Step 15760: {'lr': 0.0004052869450695776, 'samples': 3026112, 'steps': 15760, 'loss/train': 1.632610559463501} +01/27/2022 10:40:31 - INFO - codeparrot_training - Step 15761: {'lr': 0.00040527412158277744, 'samples': 3026304, 'steps': 15761, 'loss/train': 0.7356557697057724} +01/27/2022 10:40:34 - INFO - codeparrot_training - Step 15762: {'lr': 0.00040526129743083216, 'samples': 3026496, 'steps': 15762, 'loss/train': 0.627466544508934} +01/27/2022 10:40:38 - INFO - codeparrot_training - Step 15763: {'lr': 0.0004052484726137968, 'samples': 3026688, 'steps': 15763, 'loss/train': 1.2493381798267365} +01/27/2022 10:40:41 - INFO - codeparrot_training - Step 15764: {'lr': 0.00040523564713172634, 'samples': 3026880, 'steps': 15764, 'loss/train': 0.673240140080452} +01/27/2022 10:40:44 - INFO - codeparrot_training - Step 15765: {'lr': 0.0004052228209846756, 'samples': 3027072, 'steps': 15765, 'loss/train': 0.5794259458780289} +01/27/2022 10:40:47 - INFO - codeparrot_training - Step 15766: {'lr': 0.0004052099941726996, 'samples': 3027264, 'steps': 15766, 'loss/train': 1.0197400152683258} +01/27/2022 10:40:50 - INFO - codeparrot_training - Step 15767: {'lr': 0.0004051971666958533, 'samples': 3027456, 'steps': 15767, 'loss/train': 0.40519653260707855} +01/27/2022 10:40:53 - INFO - codeparrot_training - Step 15768: {'lr': 0.0004051843385541916, 'samples': 3027648, 'steps': 15768, 'loss/train': 0.778348982334137} +01/27/2022 10:40:56 - INFO - codeparrot_training - Step 15769: {'lr': 0.00040517150974776945, 'samples': 3027840, 'steps': 15769, 'loss/train': 0.9875755906105042} +01/27/2022 10:41:01 - INFO - codeparrot_training - Step 15770: {'lr': 0.00040515868027664185, 'samples': 3028032, 'steps': 15770, 'loss/train': 0.34746886789798737} +01/27/2022 10:41:04 - INFO - codeparrot_training - Step 15771: {'lr': 0.00040514585014086367, 'samples': 3028224, 'steps': 15771, 'loss/train': 0.8838028907775879} +01/27/2022 10:41:07 - INFO - codeparrot_training - Step 15772: {'lr': 0.00040513301934049005, 'samples': 3028416, 'steps': 15772, 'loss/train': 0.8548937737941742} +01/27/2022 10:41:10 - INFO - codeparrot_training - Step 15773: {'lr': 0.00040512018787557574, 'samples': 3028608, 'steps': 15773, 'loss/train': 0.8801071643829346} +01/27/2022 10:41:13 - INFO - codeparrot_training - Step 15774: {'lr': 0.0004051073557461759, 'samples': 3028800, 'steps': 15774, 'loss/train': 0.967812180519104} +01/27/2022 10:41:17 - INFO - codeparrot_training - Step 15775: {'lr': 0.00040509452295234527, 'samples': 3028992, 'steps': 15775, 'loss/train': 0.8506843149662018} +01/27/2022 10:41:20 - INFO - codeparrot_training - Step 15776: {'lr': 0.00040508168949413904, 'samples': 3029184, 'steps': 15776, 'loss/train': 0.692777082324028} +01/27/2022 10:41:23 - INFO - codeparrot_training - Step 15777: {'lr': 0.0004050688553716121, 'samples': 3029376, 'steps': 15777, 'loss/train': 0.8200711905956268} +01/27/2022 10:41:26 - INFO - codeparrot_training - Step 15778: {'lr': 0.0004050560205848194, 'samples': 3029568, 'steps': 15778, 'loss/train': 0.34250224381685257} +01/27/2022 10:41:31 - INFO - codeparrot_training - Step 15779: {'lr': 0.0004050431851338159, 'samples': 3029760, 'steps': 15779, 'loss/train': 1.7127978801727295} +01/27/2022 10:41:34 - INFO - codeparrot_training - Step 15780: {'lr': 0.00040503034901865666, 'samples': 3029952, 'steps': 15780, 'loss/train': 1.0358848571777344} +01/27/2022 10:41:37 - INFO - codeparrot_training - Step 15781: {'lr': 0.00040501751223939665, 'samples': 3030144, 'steps': 15781, 'loss/train': 0.7337314635515213} +01/27/2022 10:41:40 - INFO - codeparrot_training - Step 15782: {'lr': 0.00040500467479609084, 'samples': 3030336, 'steps': 15782, 'loss/train': 1.134171187877655} +01/27/2022 10:41:43 - INFO - codeparrot_training - Step 15783: {'lr': 0.00040499183668879415, 'samples': 3030528, 'steps': 15783, 'loss/train': 0.7894969582557678} +01/27/2022 10:41:46 - INFO - codeparrot_training - Step 15784: {'lr': 0.0004049789979175617, 'samples': 3030720, 'steps': 15784, 'loss/train': 0.7801816463470459} +01/27/2022 10:41:49 - INFO - codeparrot_training - Step 15785: {'lr': 0.00040496615848244845, 'samples': 3030912, 'steps': 15785, 'loss/train': 1.128368228673935} +01/27/2022 10:41:53 - INFO - codeparrot_training - Step 15786: {'lr': 0.00040495331838350933, 'samples': 3031104, 'steps': 15786, 'loss/train': 0.4084232300519943} +01/27/2022 10:41:57 - INFO - codeparrot_training - Step 15787: {'lr': 0.00040494047762079953, 'samples': 3031296, 'steps': 15787, 'loss/train': 1.103697806596756} +01/27/2022 10:42:00 - INFO - codeparrot_training - Step 15788: {'lr': 0.0004049276361943738, 'samples': 3031488, 'steps': 15788, 'loss/train': 1.1364565193653107} +01/27/2022 10:42:03 - INFO - codeparrot_training - Step 15789: {'lr': 0.00040491479410428735, 'samples': 3031680, 'steps': 15789, 'loss/train': 1.4359527826309204} +01/27/2022 10:42:06 - INFO - codeparrot_training - Step 15790: {'lr': 0.00040490195135059503, 'samples': 3031872, 'steps': 15790, 'loss/train': 0.6061565279960632} +01/27/2022 10:42:10 - INFO - codeparrot_training - Step 15791: {'lr': 0.000404889107933352, 'samples': 3032064, 'steps': 15791, 'loss/train': 0.290904276072979} +01/27/2022 10:42:13 - INFO - codeparrot_training - Step 15792: {'lr': 0.0004048762638526132, 'samples': 3032256, 'steps': 15792, 'loss/train': 0.8411537110805511} +01/27/2022 10:42:16 - INFO - codeparrot_training - Step 15793: {'lr': 0.0004048634191084336, 'samples': 3032448, 'steps': 15793, 'loss/train': 0.7670989036560059} +01/27/2022 10:42:19 - INFO - codeparrot_training - Step 15794: {'lr': 0.0004048505737008684, 'samples': 3032640, 'steps': 15794, 'loss/train': 0.3804671913385391} +01/27/2022 10:42:22 - INFO - codeparrot_training - Step 15795: {'lr': 0.0004048377276299724, 'samples': 3032832, 'steps': 15795, 'loss/train': 0.4056909531354904} +01/27/2022 10:42:27 - INFO - codeparrot_training - Step 15796: {'lr': 0.00040482488089580083, 'samples': 3033024, 'steps': 15796, 'loss/train': 0.9631420969963074} +01/27/2022 10:42:30 - INFO - codeparrot_training - Step 15797: {'lr': 0.00040481203349840864, 'samples': 3033216, 'steps': 15797, 'loss/train': 0.8404028713703156} +01/27/2022 10:42:34 - INFO - codeparrot_training - Step 15798: {'lr': 0.0004047991854378508, 'samples': 3033408, 'steps': 15798, 'loss/train': 0.8815054893493652} +01/27/2022 10:42:37 - INFO - codeparrot_training - Step 15799: {'lr': 0.00040478633671418244, 'samples': 3033600, 'steps': 15799, 'loss/train': 0.9557957947254181} +01/27/2022 10:42:40 - INFO - codeparrot_training - Step 15800: {'lr': 0.00040477348732745853, 'samples': 3033792, 'steps': 15800, 'loss/train': 0.4363967031240463} +01/27/2022 10:42:43 - INFO - codeparrot_training - Step 15801: {'lr': 0.00040476063727773416, 'samples': 3033984, 'steps': 15801, 'loss/train': 1.3895502984523773} +01/27/2022 10:42:46 - INFO - codeparrot_training - Step 15802: {'lr': 0.0004047477865650644, 'samples': 3034176, 'steps': 15802, 'loss/train': 0.6610544621944427} +01/27/2022 10:42:49 - INFO - codeparrot_training - Step 15803: {'lr': 0.00040473493518950414, 'samples': 3034368, 'steps': 15803, 'loss/train': 0.7326797693967819} +01/27/2022 10:42:52 - INFO - codeparrot_training - Step 15804: {'lr': 0.00040472208315110866, 'samples': 3034560, 'steps': 15804, 'loss/train': 0.7310878783464432} +01/27/2022 10:42:57 - INFO - codeparrot_training - Step 15805: {'lr': 0.0004047092304499329, 'samples': 3034752, 'steps': 15805, 'loss/train': 0.9037556648254395} +01/27/2022 10:43:00 - INFO - codeparrot_training - Step 15806: {'lr': 0.0004046963770860319, 'samples': 3034944, 'steps': 15806, 'loss/train': 0.9054333865642548} +01/27/2022 10:43:03 - INFO - codeparrot_training - Step 15807: {'lr': 0.0004046835230594608, 'samples': 3035136, 'steps': 15807, 'loss/train': 0.7851873636245728} +01/27/2022 10:43:06 - INFO - codeparrot_training - Step 15808: {'lr': 0.0004046706683702744, 'samples': 3035328, 'steps': 15808, 'loss/train': 1.018451303243637} +01/27/2022 10:43:09 - INFO - codeparrot_training - Step 15809: {'lr': 0.0004046578130185282, 'samples': 3035520, 'steps': 15809, 'loss/train': 0.9557715654373169} +01/27/2022 10:43:13 - INFO - codeparrot_training - Step 15810: {'lr': 0.00040464495700427694, 'samples': 3035712, 'steps': 15810, 'loss/train': 0.4876471906900406} +01/27/2022 10:43:16 - INFO - codeparrot_training - Step 15811: {'lr': 0.0004046321003275759, 'samples': 3035904, 'steps': 15811, 'loss/train': 0.900312602519989} +01/27/2022 10:43:19 - INFO - codeparrot_training - Step 15812: {'lr': 0.00040461924298847987, 'samples': 3036096, 'steps': 15812, 'loss/train': 0.43594689667224884} +01/27/2022 10:43:25 - INFO - codeparrot_training - Step 15813: {'lr': 0.0004046063849870442, 'samples': 3036288, 'steps': 15813, 'loss/train': 0.44287633895874023} +01/27/2022 10:43:28 - INFO - codeparrot_training - Step 15814: {'lr': 0.00040459352632332387, 'samples': 3036480, 'steps': 15814, 'loss/train': 0.8975049555301666} +01/27/2022 10:43:31 - INFO - codeparrot_training - Step 15815: {'lr': 0.0004045806669973739, 'samples': 3036672, 'steps': 15815, 'loss/train': 0.8538034558296204} +01/27/2022 10:43:34 - INFO - codeparrot_training - Step 15816: {'lr': 0.00040456780700924956, 'samples': 3036864, 'steps': 15816, 'loss/train': 1.7255841493606567} +01/27/2022 10:43:37 - INFO - codeparrot_training - Step 15817: {'lr': 0.0004045549463590057, 'samples': 3037056, 'steps': 15817, 'loss/train': 1.6329957246780396} +01/27/2022 10:43:40 - INFO - codeparrot_training - Step 15818: {'lr': 0.0004045420850466975, 'samples': 3037248, 'steps': 15818, 'loss/train': 0.5929710119962692} +01/27/2022 10:43:43 - INFO - codeparrot_training - Step 15819: {'lr': 0.00040452922307238016, 'samples': 3037440, 'steps': 15819, 'loss/train': 0.8172426223754883} +01/27/2022 10:43:47 - INFO - codeparrot_training - Step 15820: {'lr': 0.00040451636043610875, 'samples': 3037632, 'steps': 15820, 'loss/train': 0.7419150620698929} +01/27/2022 10:43:50 - INFO - codeparrot_training - Step 15821: {'lr': 0.0004045034971379382, 'samples': 3037824, 'steps': 15821, 'loss/train': 0.9539506137371063} +01/27/2022 10:43:53 - INFO - codeparrot_training - Step 15822: {'lr': 0.0004044906331779238, 'samples': 3038016, 'steps': 15822, 'loss/train': 0.0652850791811943} +01/27/2022 10:43:58 - INFO - codeparrot_training - Step 15823: {'lr': 0.00040447776855612053, 'samples': 3038208, 'steps': 15823, 'loss/train': 0.44844697415828705} +01/27/2022 10:44:01 - INFO - codeparrot_training - Step 15824: {'lr': 0.0004044649032725836, 'samples': 3038400, 'steps': 15824, 'loss/train': 0.5497914254665375} +01/27/2022 10:44:04 - INFO - codeparrot_training - Step 15825: {'lr': 0.000404452037327368, 'samples': 3038592, 'steps': 15825, 'loss/train': 0.7260218113660812} +01/27/2022 10:44:07 - INFO - codeparrot_training - Step 15826: {'lr': 0.00040443917072052906, 'samples': 3038784, 'steps': 15826, 'loss/train': 0.7065645754337311} +01/27/2022 10:44:10 - INFO - codeparrot_training - Step 15827: {'lr': 0.0004044263034521216, 'samples': 3038976, 'steps': 15827, 'loss/train': 0.4435698688030243} +01/27/2022 10:44:13 - INFO - codeparrot_training - Step 15828: {'lr': 0.000404413435522201, 'samples': 3039168, 'steps': 15828, 'loss/train': 1.0625162422657013} +01/27/2022 10:44:16 - INFO - codeparrot_training - Step 15829: {'lr': 0.00040440056693082224, 'samples': 3039360, 'steps': 15829, 'loss/train': 0.3594801798462868} +01/27/2022 10:44:20 - INFO - codeparrot_training - Step 15830: {'lr': 0.0004043876976780404, 'samples': 3039552, 'steps': 15830, 'loss/train': 0.857306957244873} +01/27/2022 10:44:23 - INFO - codeparrot_training - Step 15831: {'lr': 0.0004043748277639108, 'samples': 3039744, 'steps': 15831, 'loss/train': 0.995802104473114} +01/27/2022 10:44:27 - INFO - codeparrot_training - Step 15832: {'lr': 0.0004043619571884884, 'samples': 3039936, 'steps': 15832, 'loss/train': 0.7304923832416534} +01/27/2022 10:44:30 - INFO - codeparrot_training - Step 15833: {'lr': 0.0004043490859518284, 'samples': 3040128, 'steps': 15833, 'loss/train': 0.8303113281726837} +01/27/2022 10:44:34 - INFO - codeparrot_training - Step 15834: {'lr': 0.0004043362140539859, 'samples': 3040320, 'steps': 15834, 'loss/train': 0.7152265012264252} +01/27/2022 10:44:37 - INFO - codeparrot_training - Step 15835: {'lr': 0.00040432334149501613, 'samples': 3040512, 'steps': 15835, 'loss/train': 0.9392289519309998} +01/27/2022 10:44:40 - INFO - codeparrot_training - Step 15836: {'lr': 0.00040431046827497415, 'samples': 3040704, 'steps': 15836, 'loss/train': 0.7766838669776917} +01/27/2022 10:44:43 - INFO - codeparrot_training - Step 15837: {'lr': 0.00040429759439391513, 'samples': 3040896, 'steps': 15837, 'loss/train': 0.7847301363945007} +01/27/2022 10:44:46 - INFO - codeparrot_training - Step 15838: {'lr': 0.00040428471985189416, 'samples': 3041088, 'steps': 15838, 'loss/train': 1.1564498841762543} +01/27/2022 10:44:49 - INFO - codeparrot_training - Step 15839: {'lr': 0.0004042718446489665, 'samples': 3041280, 'steps': 15839, 'loss/train': 0.35339880734682083} +01/27/2022 10:44:54 - INFO - codeparrot_training - Step 15840: {'lr': 0.0004042589687851872, 'samples': 3041472, 'steps': 15840, 'loss/train': 0.5635401606559753} +01/27/2022 10:44:57 - INFO - codeparrot_training - Step 15841: {'lr': 0.00040424609226061146, 'samples': 3041664, 'steps': 15841, 'loss/train': 0.2023550570011139} +01/27/2022 10:45:01 - INFO - codeparrot_training - Step 15842: {'lr': 0.0004042332150752944, 'samples': 3041856, 'steps': 15842, 'loss/train': 1.5135743021965027} +01/27/2022 10:45:04 - INFO - codeparrot_training - Step 15843: {'lr': 0.0004042203372292913, 'samples': 3042048, 'steps': 15843, 'loss/train': 0.9303327798843384} +01/27/2022 10:45:07 - INFO - codeparrot_training - Step 15844: {'lr': 0.00040420745872265726, 'samples': 3042240, 'steps': 15844, 'loss/train': 1.449015587568283} +01/27/2022 10:45:10 - INFO - codeparrot_training - Step 15845: {'lr': 0.0004041945795554474, 'samples': 3042432, 'steps': 15845, 'loss/train': 0.7997302114963531} +01/27/2022 10:45:13 - INFO - codeparrot_training - Step 15846: {'lr': 0.0004041816997277169, 'samples': 3042624, 'steps': 15846, 'loss/train': 1.0357554852962494} +01/27/2022 10:45:17 - INFO - codeparrot_training - Step 15847: {'lr': 0.000404168819239521, 'samples': 3042816, 'steps': 15847, 'loss/train': 1.1277852952480316} +01/27/2022 10:45:20 - INFO - codeparrot_training - Step 15848: {'lr': 0.0004041559380909148, 'samples': 3043008, 'steps': 15848, 'loss/train': 0.5053092241287231} +01/27/2022 10:45:24 - INFO - codeparrot_training - Step 15849: {'lr': 0.00040414305628195347, 'samples': 3043200, 'steps': 15849, 'loss/train': 1.7235082983970642} +01/27/2022 10:45:27 - INFO - codeparrot_training - Step 15850: {'lr': 0.00040413017381269237, 'samples': 3043392, 'steps': 15850, 'loss/train': 0.8440417349338531} +01/27/2022 10:45:30 - INFO - codeparrot_training - Step 15851: {'lr': 0.00040411729068318635, 'samples': 3043584, 'steps': 15851, 'loss/train': 0.8654311001300812} +01/27/2022 10:45:34 - INFO - codeparrot_training - Step 15852: {'lr': 0.0004041044068934909, 'samples': 3043776, 'steps': 15852, 'loss/train': 0.9661687016487122} +01/27/2022 10:45:37 - INFO - codeparrot_training - Step 15853: {'lr': 0.00040409152244366117, 'samples': 3043968, 'steps': 15853, 'loss/train': 0.6278281956911087} +01/27/2022 10:45:40 - INFO - codeparrot_training - Step 15854: {'lr': 0.00040407863733375217, 'samples': 3044160, 'steps': 15854, 'loss/train': 1.0331066250801086} +01/27/2022 10:45:43 - INFO - codeparrot_training - Step 15855: {'lr': 0.0004040657515638193, 'samples': 3044352, 'steps': 15855, 'loss/train': 0.7883612215518951} +01/27/2022 10:45:46 - INFO - codeparrot_training - Step 15856: {'lr': 0.0004040528651339176, 'samples': 3044544, 'steps': 15856, 'loss/train': 0.6156122535467148} +01/27/2022 10:45:49 - INFO - codeparrot_training - Step 15857: {'lr': 0.00040403997804410244, 'samples': 3044736, 'steps': 15857, 'loss/train': 0.5855689197778702} +01/27/2022 10:45:54 - INFO - codeparrot_training - Step 15858: {'lr': 0.00040402709029442883, 'samples': 3044928, 'steps': 15858, 'loss/train': 0.6567723602056503} +01/27/2022 10:45:57 - INFO - codeparrot_training - Step 15859: {'lr': 0.0004040142018849521, 'samples': 3045120, 'steps': 15859, 'loss/train': 0.9871104061603546} +01/27/2022 10:46:00 - INFO - codeparrot_training - Step 15860: {'lr': 0.0004040013128157275, 'samples': 3045312, 'steps': 15860, 'loss/train': 0.8779299259185791} +01/27/2022 10:46:03 - INFO - codeparrot_training - Step 15861: {'lr': 0.0004039884230868101, 'samples': 3045504, 'steps': 15861, 'loss/train': 1.5136545896530151} +01/27/2022 10:46:06 - INFO - codeparrot_training - Step 15862: {'lr': 0.0004039755326982552, 'samples': 3045696, 'steps': 15862, 'loss/train': 0.8249252736568451} +01/27/2022 10:46:09 - INFO - codeparrot_training - Step 15863: {'lr': 0.000403962641650118, 'samples': 3045888, 'steps': 15863, 'loss/train': 1.032995492219925} +01/27/2022 10:46:12 - INFO - codeparrot_training - Step 15864: {'lr': 0.0004039497499424538, 'samples': 3046080, 'steps': 15864, 'loss/train': 1.024817705154419} +01/27/2022 10:46:16 - INFO - codeparrot_training - Step 15865: {'lr': 0.00040393685757531776, 'samples': 3046272, 'steps': 15865, 'loss/train': 0.7738977670669556} +01/27/2022 10:46:19 - INFO - codeparrot_training - Step 15866: {'lr': 0.000403923964548765, 'samples': 3046464, 'steps': 15866, 'loss/train': 0.737411767244339} +01/27/2022 10:46:24 - INFO - codeparrot_training - Step 15867: {'lr': 0.0004039110708628509, 'samples': 3046656, 'steps': 15867, 'loss/train': 0.4975927323102951} +01/27/2022 10:46:27 - INFO - codeparrot_training - Step 15868: {'lr': 0.00040389817651763073, 'samples': 3046848, 'steps': 15868, 'loss/train': 0.6608000546693802} +01/27/2022 10:46:30 - INFO - codeparrot_training - Step 15869: {'lr': 0.0004038852815131595, 'samples': 3047040, 'steps': 15869, 'loss/train': 1.15775665640831} +01/27/2022 10:46:33 - INFO - codeparrot_training - Step 15870: {'lr': 0.0004038723858494927, 'samples': 3047232, 'steps': 15870, 'loss/train': 0.8909288048744202} +01/27/2022 10:46:37 - INFO - codeparrot_training - Step 15871: {'lr': 0.00040385948952668537, 'samples': 3047424, 'steps': 15871, 'loss/train': 0.4548579901456833} +01/27/2022 10:46:40 - INFO - codeparrot_training - Step 15872: {'lr': 0.0004038465925447929, 'samples': 3047616, 'steps': 15872, 'loss/train': 1.8405178785324097} +01/27/2022 10:46:43 - INFO - codeparrot_training - Step 15873: {'lr': 0.00040383369490387043, 'samples': 3047808, 'steps': 15873, 'loss/train': 0.8154062032699585} +01/27/2022 10:46:46 - INFO - codeparrot_training - Step 15874: {'lr': 0.0004038207966039733, 'samples': 3048000, 'steps': 15874, 'loss/train': 1.0257627367973328} +01/27/2022 10:46:49 - INFO - codeparrot_training - Step 15875: {'lr': 0.00040380789764515667, 'samples': 3048192, 'steps': 15875, 'loss/train': 0.5186170041561127} +01/27/2022 10:46:53 - INFO - codeparrot_training - Step 15876: {'lr': 0.0004037949980274759, 'samples': 3048384, 'steps': 15876, 'loss/train': 0.7927306294441223} +01/27/2022 10:46:57 - INFO - codeparrot_training - Step 15877: {'lr': 0.0004037820977509862, 'samples': 3048576, 'steps': 15877, 'loss/train': 1.4380521476268768} +01/27/2022 10:47:00 - INFO - codeparrot_training - Step 15878: {'lr': 0.00040376919681574285, 'samples': 3048768, 'steps': 15878, 'loss/train': 1.3119031190872192} +01/27/2022 10:47:03 - INFO - codeparrot_training - Step 15879: {'lr': 0.000403756295221801, 'samples': 3048960, 'steps': 15879, 'loss/train': 0.5754084438085556} +01/27/2022 10:47:06 - INFO - codeparrot_training - Step 15880: {'lr': 0.00040374339296921606, 'samples': 3049152, 'steps': 15880, 'loss/train': 0.837433934211731} +01/27/2022 10:47:09 - INFO - codeparrot_training - Step 15881: {'lr': 0.00040373049005804323, 'samples': 3049344, 'steps': 15881, 'loss/train': 0.7528701424598694} +01/27/2022 10:47:12 - INFO - codeparrot_training - Step 15882: {'lr': 0.00040371758648833776, 'samples': 3049536, 'steps': 15882, 'loss/train': 0.8456114530563354} +01/27/2022 10:47:15 - INFO - codeparrot_training - Step 15883: {'lr': 0.00040370468226015507, 'samples': 3049728, 'steps': 15883, 'loss/train': 1.1196337938308716} +01/27/2022 10:47:20 - INFO - codeparrot_training - Step 15884: {'lr': 0.0004036917773735502, 'samples': 3049920, 'steps': 15884, 'loss/train': 0.09344386495649815} +01/27/2022 10:47:23 - INFO - codeparrot_training - Step 15885: {'lr': 0.00040367887182857866, 'samples': 3050112, 'steps': 15885, 'loss/train': 0.16653797775506973} +01/27/2022 10:47:26 - INFO - codeparrot_training - Step 15886: {'lr': 0.00040366596562529554, 'samples': 3050304, 'steps': 15886, 'loss/train': 0.8097220659255981} +01/27/2022 10:47:29 - INFO - codeparrot_training - Step 15887: {'lr': 0.00040365305876375636, 'samples': 3050496, 'steps': 15887, 'loss/train': 1.0628576874732971} +01/27/2022 10:47:32 - INFO - codeparrot_training - Step 15888: {'lr': 0.0004036401512440161, 'samples': 3050688, 'steps': 15888, 'loss/train': 0.9778043031692505} +01/27/2022 10:47:36 - INFO - codeparrot_training - Step 15889: {'lr': 0.0004036272430661303, 'samples': 3050880, 'steps': 15889, 'loss/train': 0.8734466135501862} +01/27/2022 10:47:39 - INFO - codeparrot_training - Step 15890: {'lr': 0.0004036143342301542, 'samples': 3051072, 'steps': 15890, 'loss/train': 1.4610282182693481} +01/27/2022 10:47:42 - INFO - codeparrot_training - Step 15891: {'lr': 0.000403601424736143, 'samples': 3051264, 'steps': 15891, 'loss/train': 1.32945317029953} +01/27/2022 10:47:45 - INFO - codeparrot_training - Step 15892: {'lr': 0.0004035885145841521, 'samples': 3051456, 'steps': 15892, 'loss/train': 0.22864260524511337} +01/27/2022 10:47:50 - INFO - codeparrot_training - Step 15893: {'lr': 0.00040357560377423675, 'samples': 3051648, 'steps': 15893, 'loss/train': 0.6869247257709503} +01/27/2022 10:47:53 - INFO - codeparrot_training - Step 15894: {'lr': 0.0004035626923064524, 'samples': 3051840, 'steps': 15894, 'loss/train': 0.8503405451774597} +01/27/2022 10:47:56 - INFO - codeparrot_training - Step 15895: {'lr': 0.00040354978018085407, 'samples': 3052032, 'steps': 15895, 'loss/train': 0.8380831182003021} +01/27/2022 10:48:00 - INFO - codeparrot_training - Step 15896: {'lr': 0.00040353686739749733, 'samples': 3052224, 'steps': 15896, 'loss/train': 1.433519035577774} +01/27/2022 10:48:03 - INFO - codeparrot_training - Step 15897: {'lr': 0.00040352395395643737, 'samples': 3052416, 'steps': 15897, 'loss/train': 1.0187241733074188} +01/27/2022 10:48:06 - INFO - codeparrot_training - Step 15898: {'lr': 0.00040351103985772964, 'samples': 3052608, 'steps': 15898, 'loss/train': 0.711328312754631} +01/27/2022 10:48:09 - INFO - codeparrot_training - Step 15899: {'lr': 0.00040349812510142923, 'samples': 3052800, 'steps': 15899, 'loss/train': 1.03606715798378} +01/27/2022 10:48:12 - INFO - codeparrot_training - Step 15900: {'lr': 0.0004034852096875916, 'samples': 3052992, 'steps': 15900, 'loss/train': 0.8503400087356567} +01/27/2022 10:48:15 - INFO - codeparrot_training - Step 15901: {'lr': 0.0004034722936162721, 'samples': 3053184, 'steps': 15901, 'loss/train': 0.8401461839675903} +01/27/2022 10:48:20 - INFO - codeparrot_training - Step 15902: {'lr': 0.00040345937688752607, 'samples': 3053376, 'steps': 15902, 'loss/train': 0.9001806378364563} +01/27/2022 10:48:23 - INFO - codeparrot_training - Step 15903: {'lr': 0.0004034464595014088, 'samples': 3053568, 'steps': 15903, 'loss/train': 1.013807773590088} +01/27/2022 10:48:26 - INFO - codeparrot_training - Step 15904: {'lr': 0.00040343354145797554, 'samples': 3053760, 'steps': 15904, 'loss/train': 0.7260244488716125} +01/27/2022 10:48:29 - INFO - codeparrot_training - Step 15905: {'lr': 0.0004034206227572818, 'samples': 3053952, 'steps': 15905, 'loss/train': 0.5656226724386215} +01/27/2022 10:48:32 - INFO - codeparrot_training - Step 15906: {'lr': 0.0004034077033993828, 'samples': 3054144, 'steps': 15906, 'loss/train': 0.8087723851203918} +01/27/2022 10:48:35 - INFO - codeparrot_training - Step 15907: {'lr': 0.00040339478338433386, 'samples': 3054336, 'steps': 15907, 'loss/train': 0.606577679514885} +01/27/2022 10:48:39 - INFO - codeparrot_training - Step 15908: {'lr': 0.0004033818627121904, 'samples': 3054528, 'steps': 15908, 'loss/train': 0.5134348422288895} +01/27/2022 10:48:42 - INFO - codeparrot_training - Step 15909: {'lr': 0.00040336894138300777, 'samples': 3054720, 'steps': 15909, 'loss/train': 0.7509387731552124} +01/27/2022 10:48:46 - INFO - codeparrot_training - Step 15910: {'lr': 0.0004033560193968413, 'samples': 3054912, 'steps': 15910, 'loss/train': 0.8957132399082184} +01/27/2022 10:48:49 - INFO - codeparrot_training - Step 15911: {'lr': 0.00040334309675374636, 'samples': 3055104, 'steps': 15911, 'loss/train': 0.29854072630405426} +01/27/2022 10:48:53 - INFO - codeparrot_training - Step 15912: {'lr': 0.0004033301734537782, 'samples': 3055296, 'steps': 15912, 'loss/train': 0.595838651061058} +01/27/2022 10:48:56 - INFO - codeparrot_training - Step 15913: {'lr': 0.0004033172494969923, 'samples': 3055488, 'steps': 15913, 'loss/train': 1.4226442873477936} +01/27/2022 10:48:59 - INFO - codeparrot_training - Step 15914: {'lr': 0.000403304324883444, 'samples': 3055680, 'steps': 15914, 'loss/train': 1.1866225898265839} +01/27/2022 10:49:02 - INFO - codeparrot_training - Step 15915: {'lr': 0.00040329139961318863, 'samples': 3055872, 'steps': 15915, 'loss/train': 0.6561603248119354} +01/27/2022 10:49:05 - INFO - codeparrot_training - Step 15916: {'lr': 0.00040327847368628163, 'samples': 3056064, 'steps': 15916, 'loss/train': 0.8509580790996552} +01/27/2022 10:49:08 - INFO - codeparrot_training - Step 15917: {'lr': 0.0004032655471027783, 'samples': 3056256, 'steps': 15917, 'loss/train': 0.4859050065279007} +01/27/2022 10:49:11 - INFO - codeparrot_training - Step 15918: {'lr': 0.000403252619862734, 'samples': 3056448, 'steps': 15918, 'loss/train': 0.5695238560438156} +01/27/2022 10:49:17 - INFO - codeparrot_training - Step 15919: {'lr': 0.0004032396919662041, 'samples': 3056640, 'steps': 15919, 'loss/train': 1.5581914186477661} +01/27/2022 10:49:20 - INFO - codeparrot_training - Step 15920: {'lr': 0.00040322676341324415, 'samples': 3056832, 'steps': 15920, 'loss/train': 0.9792717397212982} +01/27/2022 10:49:23 - INFO - codeparrot_training - Step 15921: {'lr': 0.0004032138342039093, 'samples': 3057024, 'steps': 15921, 'loss/train': 0.03607937414199114} +01/27/2022 10:49:26 - INFO - codeparrot_training - Step 15922: {'lr': 0.0004032009043382551, 'samples': 3057216, 'steps': 15922, 'loss/train': 1.1417209804058075} +01/27/2022 10:49:29 - INFO - codeparrot_training - Step 15923: {'lr': 0.0004031879738163368, 'samples': 3057408, 'steps': 15923, 'loss/train': 0.9061854779720306} +01/27/2022 10:49:32 - INFO - codeparrot_training - Step 15924: {'lr': 0.00040317504263820994, 'samples': 3057600, 'steps': 15924, 'loss/train': 1.0217056274414062} +01/27/2022 10:49:35 - INFO - codeparrot_training - Step 15925: {'lr': 0.0004031621108039298, 'samples': 3057792, 'steps': 15925, 'loss/train': 0.7702926993370056} +01/27/2022 10:49:39 - INFO - codeparrot_training - Step 15926: {'lr': 0.0004031491783135518, 'samples': 3057984, 'steps': 15926, 'loss/train': 0.46267475187778473} +01/27/2022 10:49:42 - INFO - codeparrot_training - Step 15927: {'lr': 0.0004031362451671314, 'samples': 3058176, 'steps': 15927, 'loss/train': 0.6191606372594833} +01/27/2022 10:49:46 - INFO - codeparrot_training - Step 15928: {'lr': 0.00040312331136472385, 'samples': 3058368, 'steps': 15928, 'loss/train': 0.7884693145751953} +01/27/2022 10:49:49 - INFO - codeparrot_training - Step 15929: {'lr': 0.00040311037690638477, 'samples': 3058560, 'steps': 15929, 'loss/train': 1.034155011177063} +01/27/2022 10:49:53 - INFO - codeparrot_training - Step 15930: {'lr': 0.00040309744179216936, 'samples': 3058752, 'steps': 15930, 'loss/train': 0.7509830296039581} +01/27/2022 10:49:56 - INFO - codeparrot_training - Step 15931: {'lr': 0.0004030845060221332, 'samples': 3058944, 'steps': 15931, 'loss/train': 0.5110121816396713} +01/27/2022 10:49:59 - INFO - codeparrot_training - Step 15932: {'lr': 0.00040307156959633154, 'samples': 3059136, 'steps': 15932, 'loss/train': 0.572925791144371} +01/27/2022 10:50:02 - INFO - codeparrot_training - Step 15933: {'lr': 0.00040305863251482, 'samples': 3059328, 'steps': 15933, 'loss/train': 0.8698444962501526} +01/27/2022 10:50:05 - INFO - codeparrot_training - Step 15934: {'lr': 0.00040304569477765375, 'samples': 3059520, 'steps': 15934, 'loss/train': 0.621804803609848} +01/27/2022 10:50:08 - INFO - codeparrot_training - Step 15935: {'lr': 0.0004030327563848885, 'samples': 3059712, 'steps': 15935, 'loss/train': 0.5994575768709183} +01/27/2022 10:50:12 - INFO - codeparrot_training - Step 15936: {'lr': 0.00040301981733657934, 'samples': 3059904, 'steps': 15936, 'loss/train': 0.7923910617828369} +01/27/2022 10:50:16 - INFO - codeparrot_training - Step 15937: {'lr': 0.00040300687763278196, 'samples': 3060096, 'steps': 15937, 'loss/train': 0.9408222734928131} +01/27/2022 10:50:19 - INFO - codeparrot_training - Step 15938: {'lr': 0.0004029939372735517, 'samples': 3060288, 'steps': 15938, 'loss/train': 1.0263008773326874} +01/27/2022 10:50:22 - INFO - codeparrot_training - Step 15939: {'lr': 0.000402980996258944, 'samples': 3060480, 'steps': 15939, 'loss/train': 0.7715176641941071} +01/27/2022 10:50:25 - INFO - codeparrot_training - Step 15940: {'lr': 0.00040296805458901427, 'samples': 3060672, 'steps': 15940, 'loss/train': 0.8964411914348602} +01/27/2022 10:50:29 - INFO - codeparrot_training - Step 15941: {'lr': 0.0004029551122638179, 'samples': 3060864, 'steps': 15941, 'loss/train': 0.034213246777653694} +01/27/2022 10:50:32 - INFO - codeparrot_training - Step 15942: {'lr': 0.0004029421692834105, 'samples': 3061056, 'steps': 15942, 'loss/train': 0.7306340932846069} +01/27/2022 10:50:35 - INFO - codeparrot_training - Step 15943: {'lr': 0.0004029292256478474, 'samples': 3061248, 'steps': 15943, 'loss/train': 0.914590448141098} +01/27/2022 10:50:38 - INFO - codeparrot_training - Step 15944: {'lr': 0.00040291628135718404, 'samples': 3061440, 'steps': 15944, 'loss/train': 0.29969096928834915} +01/27/2022 10:50:41 - INFO - codeparrot_training - Step 15945: {'lr': 0.0004029033364114759, 'samples': 3061632, 'steps': 15945, 'loss/train': 0.967570424079895} +01/27/2022 10:50:46 - INFO - codeparrot_training - Step 15946: {'lr': 0.00040289039081077837, 'samples': 3061824, 'steps': 15946, 'loss/train': 0.06510771811008453} +01/27/2022 10:50:49 - INFO - codeparrot_training - Step 15947: {'lr': 0.00040287744455514703, 'samples': 3062016, 'steps': 15947, 'loss/train': 0.5416173934936523} +01/27/2022 10:50:53 - INFO - codeparrot_training - Step 15948: {'lr': 0.00040286449764463715, 'samples': 3062208, 'steps': 15948, 'loss/train': 1.2154988944530487} +01/27/2022 10:50:56 - INFO - codeparrot_training - Step 15949: {'lr': 0.0004028515500793044, 'samples': 3062400, 'steps': 15949, 'loss/train': 0.6943444758653641} +01/27/2022 10:50:59 - INFO - codeparrot_training - Step 15950: {'lr': 0.0004028386018592041, 'samples': 3062592, 'steps': 15950, 'loss/train': 0.7367964684963226} +01/27/2022 10:51:02 - INFO - codeparrot_training - Step 15951: {'lr': 0.0004028256529843918, 'samples': 3062784, 'steps': 15951, 'loss/train': 0.1951119303703308} +01/27/2022 10:51:05 - INFO - codeparrot_training - Step 15952: {'lr': 0.00040281270345492295, 'samples': 3062976, 'steps': 15952, 'loss/train': 0.7404391765594482} +01/27/2022 10:51:08 - INFO - codeparrot_training - Step 15953: {'lr': 0.00040279975327085294, 'samples': 3063168, 'steps': 15953, 'loss/train': 1.019188016653061} +01/27/2022 10:51:13 - INFO - codeparrot_training - Step 15954: {'lr': 0.00040278680243223733, 'samples': 3063360, 'steps': 15954, 'loss/train': 1.4263292849063873} +01/27/2022 10:51:16 - INFO - codeparrot_training - Step 15955: {'lr': 0.00040277385093913154, 'samples': 3063552, 'steps': 15955, 'loss/train': 0.7635740339756012} +01/27/2022 10:51:19 - INFO - codeparrot_training - Step 15956: {'lr': 0.0004027608987915912, 'samples': 3063744, 'steps': 15956, 'loss/train': 0.9450821578502655} +01/27/2022 10:51:22 - INFO - codeparrot_training - Step 15957: {'lr': 0.0004027479459896716, 'samples': 3063936, 'steps': 15957, 'loss/train': 0.877367377281189} +01/27/2022 10:51:25 - INFO - codeparrot_training - Step 15958: {'lr': 0.0004027349925334282, 'samples': 3064128, 'steps': 15958, 'loss/train': 0.9186590015888214} +01/27/2022 10:51:29 - INFO - codeparrot_training - Step 15959: {'lr': 0.00040272203842291676, 'samples': 3064320, 'steps': 15959, 'loss/train': 0.5872037708759308} +01/27/2022 10:51:32 - INFO - codeparrot_training - Step 15960: {'lr': 0.00040270908365819247, 'samples': 3064512, 'steps': 15960, 'loss/train': 0.741337850689888} +01/27/2022 10:51:35 - INFO - codeparrot_training - Step 15961: {'lr': 0.000402696128239311, 'samples': 3064704, 'steps': 15961, 'loss/train': 0.7080902606248856} +01/27/2022 10:51:38 - INFO - codeparrot_training - Step 15962: {'lr': 0.00040268317216632783, 'samples': 3064896, 'steps': 15962, 'loss/train': 1.0211680233478546} +01/27/2022 10:51:43 - INFO - codeparrot_training - Step 15963: {'lr': 0.0004026702154392984, 'samples': 3065088, 'steps': 15963, 'loss/train': 1.3270134329795837} +01/27/2022 10:51:46 - INFO - codeparrot_training - Step 15964: {'lr': 0.0004026572580582783, 'samples': 3065280, 'steps': 15964, 'loss/train': 0.8921737968921661} +01/27/2022 10:51:49 - INFO - codeparrot_training - Step 15965: {'lr': 0.000402644300023323, 'samples': 3065472, 'steps': 15965, 'loss/train': 1.315031111240387} +01/27/2022 10:51:53 - INFO - codeparrot_training - Step 15966: {'lr': 0.0004026313413344879, 'samples': 3065664, 'steps': 15966, 'loss/train': 0.5468924939632416} +01/27/2022 10:51:56 - INFO - codeparrot_training - Step 15967: {'lr': 0.0004026183819918286, 'samples': 3065856, 'steps': 15967, 'loss/train': 0.7950821220874786} +01/27/2022 10:51:59 - INFO - codeparrot_training - Step 15968: {'lr': 0.00040260542199540064, 'samples': 3066048, 'steps': 15968, 'loss/train': 1.3473446667194366} +01/27/2022 10:52:02 - INFO - codeparrot_training - Step 15969: {'lr': 0.00040259246134525953, 'samples': 3066240, 'steps': 15969, 'loss/train': 0.5939890891313553} +01/27/2022 10:52:05 - INFO - codeparrot_training - Step 15970: {'lr': 0.0004025795000414608, 'samples': 3066432, 'steps': 15970, 'loss/train': 1.0007121562957764} +01/27/2022 10:52:08 - INFO - codeparrot_training - Step 15971: {'lr': 0.0004025665380840599, 'samples': 3066624, 'steps': 15971, 'loss/train': 0.45434877276420593} +01/27/2022 10:52:13 - INFO - codeparrot_training - Step 15972: {'lr': 0.00040255357547311235, 'samples': 3066816, 'steps': 15972, 'loss/train': 0.8398591876029968} +01/27/2022 10:52:16 - INFO - codeparrot_training - Step 15973: {'lr': 0.0004025406122086738, 'samples': 3067008, 'steps': 15973, 'loss/train': 0.606710895895958} +01/27/2022 10:52:19 - INFO - codeparrot_training - Step 15974: {'lr': 0.0004025276482907996, 'samples': 3067200, 'steps': 15974, 'loss/train': 0.8103033900260925} +01/27/2022 10:52:22 - INFO - codeparrot_training - Step 15975: {'lr': 0.0004025146837195455, 'samples': 3067392, 'steps': 15975, 'loss/train': 0.8858403861522675} +01/27/2022 10:52:25 - INFO - codeparrot_training - Step 15976: {'lr': 0.00040250171849496685, 'samples': 3067584, 'steps': 15976, 'loss/train': 0.7394555658102036} +01/27/2022 10:52:28 - INFO - codeparrot_training - Step 15977: {'lr': 0.0004024887526171193, 'samples': 3067776, 'steps': 15977, 'loss/train': 1.1499018967151642} +01/27/2022 10:52:31 - INFO - codeparrot_training - Step 15978: {'lr': 0.0004024757860860584, 'samples': 3067968, 'steps': 15978, 'loss/train': 0.988450437784195} +01/27/2022 10:52:35 - INFO - codeparrot_training - Step 15979: {'lr': 0.00040246281890183954, 'samples': 3068160, 'steps': 15979, 'loss/train': 0.8495365083217621} +01/27/2022 10:52:38 - INFO - codeparrot_training - Step 15980: {'lr': 0.0004024498510645185, 'samples': 3068352, 'steps': 15980, 'loss/train': 0.8621210753917694} +01/27/2022 10:52:42 - INFO - codeparrot_training - Step 15981: {'lr': 0.00040243688257415064, 'samples': 3068544, 'steps': 15981, 'loss/train': 0.7319075167179108} +01/27/2022 10:52:46 - INFO - codeparrot_training - Step 15982: {'lr': 0.00040242391343079157, 'samples': 3068736, 'steps': 15982, 'loss/train': 0.6308471113443375} +01/27/2022 10:52:49 - INFO - codeparrot_training - Step 15983: {'lr': 0.00040241094363449684, 'samples': 3068928, 'steps': 15983, 'loss/train': 0.25226589292287827} +01/27/2022 10:52:52 - INFO - codeparrot_training - Step 15984: {'lr': 0.000402397973185322, 'samples': 3069120, 'steps': 15984, 'loss/train': 0.6666531264781952} +01/27/2022 10:52:55 - INFO - codeparrot_training - Step 15985: {'lr': 0.0004023850020833227, 'samples': 3069312, 'steps': 15985, 'loss/train': 0.8800175786018372} +01/27/2022 10:52:58 - INFO - codeparrot_training - Step 15986: {'lr': 0.00040237203032855446, 'samples': 3069504, 'steps': 15986, 'loss/train': 0.5587428063154221} +01/27/2022 10:53:01 - INFO - codeparrot_training - Step 15987: {'lr': 0.00040235905792107275, 'samples': 3069696, 'steps': 15987, 'loss/train': 0.8682463467121124} +01/27/2022 10:53:04 - INFO - codeparrot_training - Step 15988: {'lr': 0.00040234608486093326, 'samples': 3069888, 'steps': 15988, 'loss/train': 1.115193396806717} +01/27/2022 10:53:09 - INFO - codeparrot_training - Step 15989: {'lr': 0.00040233311114819156, 'samples': 3070080, 'steps': 15989, 'loss/train': 0.8071020841598511} +01/27/2022 10:53:12 - INFO - codeparrot_training - Step 15990: {'lr': 0.00040232013678290316, 'samples': 3070272, 'steps': 15990, 'loss/train': 1.0783033668994904} +01/27/2022 10:53:15 - INFO - codeparrot_training - Step 15991: {'lr': 0.0004023071617651236, 'samples': 3070464, 'steps': 15991, 'loss/train': 0.9162822961807251} +01/27/2022 10:53:18 - INFO - codeparrot_training - Step 15992: {'lr': 0.0004022941860949085, 'samples': 3070656, 'steps': 15992, 'loss/train': 0.7560935318470001} +01/27/2022 10:53:22 - INFO - codeparrot_training - Step 15993: {'lr': 0.00040228120977231355, 'samples': 3070848, 'steps': 15993, 'loss/train': 1.331439346075058} +01/27/2022 10:53:25 - INFO - codeparrot_training - Step 15994: {'lr': 0.00040226823279739427, 'samples': 3071040, 'steps': 15994, 'loss/train': 0.1973893716931343} +01/27/2022 10:53:28 - INFO - codeparrot_training - Step 15995: {'lr': 0.00040225525517020616, 'samples': 3071232, 'steps': 15995, 'loss/train': 0.711289644241333} +01/27/2022 10:53:31 - INFO - codeparrot_training - Step 15996: {'lr': 0.0004022422768908049, 'samples': 3071424, 'steps': 15996, 'loss/train': 0.6317132413387299} +01/27/2022 10:53:34 - INFO - codeparrot_training - Step 15997: {'lr': 0.00040222929795924613, 'samples': 3071616, 'steps': 15997, 'loss/train': 0.6880833953619003} +01/27/2022 10:53:39 - INFO - codeparrot_training - Step 15998: {'lr': 0.0004022163183755853, 'samples': 3071808, 'steps': 15998, 'loss/train': 1.232497125864029} +01/27/2022 10:53:43 - INFO - codeparrot_training - Step 15999: {'lr': 0.0004022033381398781, 'samples': 3072000, 'steps': 15999, 'loss/train': 0.9505569040775299} +01/27/2022 10:53:43 - INFO - codeparrot_training - Evaluating and saving model checkpoint