{ "best_metric": 0.10381071222883471, "best_model_checkpoint": "d:\\\\whisper-medium-pt-cv19-fleurs2-lr\\checkpoint-10000", "epoch": 11.441647597254004, "eval_steps": 5000, "global_step": 25000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.011441647597254004, "grad_norm": 17.1010799407959, "learning_rate": 2.875e-08, "loss": 0.7462, "step": 25 }, { "epoch": 0.02288329519450801, "grad_norm": 36.57398986816406, "learning_rate": 5.8750000000000007e-08, "loss": 1.2736, "step": 50 }, { "epoch": 0.034324942791762014, "grad_norm": 14.23887825012207, "learning_rate": 9e-08, "loss": 0.7475, "step": 75 }, { "epoch": 0.04576659038901602, "grad_norm": 37.83841323852539, "learning_rate": 1.2125e-07, "loss": 1.2338, "step": 100 }, { "epoch": 0.057208237986270026, "grad_norm": 13.831230163574219, "learning_rate": 1.5250000000000002e-07, "loss": 0.6932, "step": 125 }, { "epoch": 0.06864988558352403, "grad_norm": 33.25807189941406, "learning_rate": 1.8375000000000001e-07, "loss": 1.1706, "step": 150 }, { "epoch": 0.08009153318077804, "grad_norm": 15.954742431640625, "learning_rate": 2.15e-07, "loss": 0.5765, "step": 175 }, { "epoch": 0.09153318077803203, "grad_norm": 28.214441299438477, "learning_rate": 2.4624999999999997e-07, "loss": 0.7797, "step": 200 }, { "epoch": 0.10297482837528604, "grad_norm": 9.24880313873291, "learning_rate": 2.7750000000000004e-07, "loss": 0.317, "step": 225 }, { "epoch": 0.11441647597254005, "grad_norm": 15.88664722442627, "learning_rate": 3.0875e-07, "loss": 0.3906, "step": 250 }, { "epoch": 0.12585812356979406, "grad_norm": 7.954398155212402, "learning_rate": 3.4e-07, "loss": 0.2576, "step": 275 }, { "epoch": 0.13729977116704806, "grad_norm": 20.36141586303711, "learning_rate": 3.7125000000000005e-07, "loss": 0.3185, "step": 300 }, { "epoch": 0.14874141876430205, "grad_norm": 5.76043176651001, "learning_rate": 4.025e-07, "loss": 0.1886, "step": 325 }, { "epoch": 0.16018306636155608, "grad_norm": 24.122772216796875, "learning_rate": 4.3375000000000003e-07, "loss": 0.2623, "step": 350 }, { "epoch": 0.17162471395881007, "grad_norm": 8.021683692932129, "learning_rate": 4.65e-07, "loss": 0.2107, "step": 375 }, { "epoch": 0.18306636155606407, "grad_norm": 20.607276916503906, "learning_rate": 4.9625e-07, "loss": 0.2821, "step": 400 }, { "epoch": 0.1945080091533181, "grad_norm": 9.00802230834961, "learning_rate": 5.275e-07, "loss": 0.1848, "step": 425 }, { "epoch": 0.20594965675057209, "grad_norm": 16.48225975036621, "learning_rate": 5.587499999999999e-07, "loss": 0.2534, "step": 450 }, { "epoch": 0.21739130434782608, "grad_norm": 7.142576217651367, "learning_rate": 5.9e-07, "loss": 0.1887, "step": 475 }, { "epoch": 0.2288329519450801, "grad_norm": 13.774003982543945, "learning_rate": 6.212500000000001e-07, "loss": 0.2385, "step": 500 }, { "epoch": 0.2402745995423341, "grad_norm": 9.393098831176758, "learning_rate": 6.525000000000001e-07, "loss": 0.1894, "step": 525 }, { "epoch": 0.2517162471395881, "grad_norm": 17.331106185913086, "learning_rate": 6.8375e-07, "loss": 0.2414, "step": 550 }, { "epoch": 0.2631578947368421, "grad_norm": 6.325589656829834, "learning_rate": 7.15e-07, "loss": 0.1827, "step": 575 }, { "epoch": 0.2745995423340961, "grad_norm": 19.494165420532227, "learning_rate": 7.462500000000001e-07, "loss": 0.2265, "step": 600 }, { "epoch": 0.28604118993135014, "grad_norm": 8.660221099853516, "learning_rate": 7.775e-07, "loss": 0.2014, "step": 625 }, { "epoch": 0.2974828375286041, "grad_norm": 16.032672882080078, "learning_rate": 8.0875e-07, "loss": 0.2182, "step": 650 }, { "epoch": 0.30892448512585813, "grad_norm": 9.232745170593262, "learning_rate": 8.4e-07, "loss": 0.1863, "step": 675 }, { "epoch": 0.32036613272311215, "grad_norm": 17.713306427001953, "learning_rate": 8.7125e-07, "loss": 0.2344, "step": 700 }, { "epoch": 0.3318077803203661, "grad_norm": 6.892812728881836, "learning_rate": 9.025e-07, "loss": 0.1698, "step": 725 }, { "epoch": 0.34324942791762014, "grad_norm": 14.548686981201172, "learning_rate": 9.337500000000001e-07, "loss": 0.2197, "step": 750 }, { "epoch": 0.35469107551487417, "grad_norm": 6.604827404022217, "learning_rate": 9.65e-07, "loss": 0.1939, "step": 775 }, { "epoch": 0.36613272311212813, "grad_norm": 12.459800720214844, "learning_rate": 9.9625e-07, "loss": 0.2181, "step": 800 }, { "epoch": 0.37757437070938216, "grad_norm": 6.9460248947143555, "learning_rate": 1.0275e-06, "loss": 0.1656, "step": 825 }, { "epoch": 0.3890160183066362, "grad_norm": 14.406033515930176, "learning_rate": 1.05875e-06, "loss": 0.2054, "step": 850 }, { "epoch": 0.40045766590389015, "grad_norm": 4.875965118408203, "learning_rate": 1.0900000000000002e-06, "loss": 0.1801, "step": 875 }, { "epoch": 0.41189931350114417, "grad_norm": 10.513298988342285, "learning_rate": 1.12125e-06, "loss": 0.2108, "step": 900 }, { "epoch": 0.4233409610983982, "grad_norm": 6.491100311279297, "learning_rate": 1.1525000000000002e-06, "loss": 0.1719, "step": 925 }, { "epoch": 0.43478260869565216, "grad_norm": 17.76156997680664, "learning_rate": 1.18375e-06, "loss": 0.1924, "step": 950 }, { "epoch": 0.4462242562929062, "grad_norm": 4.623416900634766, "learning_rate": 1.215e-06, "loss": 0.1684, "step": 975 }, { "epoch": 0.4576659038901602, "grad_norm": 15.875606536865234, "learning_rate": 1.24625e-06, "loss": 0.1971, "step": 1000 }, { "epoch": 0.4691075514874142, "grad_norm": 5.360893726348877, "learning_rate": 1.2775e-06, "loss": 0.1661, "step": 1025 }, { "epoch": 0.4805491990846682, "grad_norm": 11.622519493103027, "learning_rate": 1.3087500000000002e-06, "loss": 0.1868, "step": 1050 }, { "epoch": 0.4919908466819222, "grad_norm": 8.163434982299805, "learning_rate": 1.34e-06, "loss": 0.1643, "step": 1075 }, { "epoch": 0.5034324942791762, "grad_norm": 14.534429550170898, "learning_rate": 1.3712500000000002e-06, "loss": 0.1727, "step": 1100 }, { "epoch": 0.5148741418764302, "grad_norm": 9.445674896240234, "learning_rate": 1.4025e-06, "loss": 0.1659, "step": 1125 }, { "epoch": 0.5263157894736842, "grad_norm": 17.26341438293457, "learning_rate": 1.43375e-06, "loss": 0.1929, "step": 1150 }, { "epoch": 0.5377574370709383, "grad_norm": 4.513171672821045, "learning_rate": 1.465e-06, "loss": 0.169, "step": 1175 }, { "epoch": 0.5491990846681922, "grad_norm": 10.165334701538086, "learning_rate": 1.49625e-06, "loss": 0.1726, "step": 1200 }, { "epoch": 0.5606407322654462, "grad_norm": 4.283674716949463, "learning_rate": 1.5275000000000002e-06, "loss": 0.1672, "step": 1225 }, { "epoch": 0.5720823798627003, "grad_norm": 14.090361595153809, "learning_rate": 1.5587500000000001e-06, "loss": 0.181, "step": 1250 }, { "epoch": 0.5835240274599542, "grad_norm": 6.832610130310059, "learning_rate": 1.5900000000000002e-06, "loss": 0.1542, "step": 1275 }, { "epoch": 0.5949656750572082, "grad_norm": 14.610993385314941, "learning_rate": 1.6212500000000001e-06, "loss": 0.2017, "step": 1300 }, { "epoch": 0.6064073226544623, "grad_norm": 6.391531467437744, "learning_rate": 1.6525000000000003e-06, "loss": 0.1752, "step": 1325 }, { "epoch": 0.6178489702517163, "grad_norm": 15.447869300842285, "learning_rate": 1.68375e-06, "loss": 0.1862, "step": 1350 }, { "epoch": 0.6292906178489702, "grad_norm": 10.263774871826172, "learning_rate": 1.7149999999999999e-06, "loss": 0.1728, "step": 1375 }, { "epoch": 0.6407322654462243, "grad_norm": 9.934491157531738, "learning_rate": 1.74625e-06, "loss": 0.162, "step": 1400 }, { "epoch": 0.6521739130434783, "grad_norm": 10.328812599182129, "learning_rate": 1.7775e-06, "loss": 0.1635, "step": 1425 }, { "epoch": 0.6636155606407322, "grad_norm": 16.462087631225586, "learning_rate": 1.80875e-06, "loss": 0.1785, "step": 1450 }, { "epoch": 0.6750572082379863, "grad_norm": 6.4470534324646, "learning_rate": 1.84e-06, "loss": 0.1706, "step": 1475 }, { "epoch": 0.6864988558352403, "grad_norm": 14.632269859313965, "learning_rate": 1.87125e-06, "loss": 0.1897, "step": 1500 }, { "epoch": 0.6979405034324943, "grad_norm": 5.021676063537598, "learning_rate": 1.9025000000000002e-06, "loss": 0.1711, "step": 1525 }, { "epoch": 0.7093821510297483, "grad_norm": 14.754373550415039, "learning_rate": 1.9337500000000003e-06, "loss": 0.2165, "step": 1550 }, { "epoch": 0.7208237986270023, "grad_norm": 8.087540626525879, "learning_rate": 1.9650000000000002e-06, "loss": 0.1793, "step": 1575 }, { "epoch": 0.7322654462242563, "grad_norm": 13.461847305297852, "learning_rate": 1.99625e-06, "loss": 0.176, "step": 1600 }, { "epoch": 0.7437070938215103, "grad_norm": 6.565804958343506, "learning_rate": 2.0275e-06, "loss": 0.1696, "step": 1625 }, { "epoch": 0.7551487414187643, "grad_norm": 16.43035888671875, "learning_rate": 2.0587500000000004e-06, "loss": 0.1796, "step": 1650 }, { "epoch": 0.7665903890160183, "grad_norm": 7.316359996795654, "learning_rate": 2.09e-06, "loss": 0.1565, "step": 1675 }, { "epoch": 0.7780320366132724, "grad_norm": 14.20641803741455, "learning_rate": 2.12125e-06, "loss": 0.1793, "step": 1700 }, { "epoch": 0.7894736842105263, "grad_norm": 5.601787090301514, "learning_rate": 2.1525e-06, "loss": 0.1518, "step": 1725 }, { "epoch": 0.8009153318077803, "grad_norm": 13.47598648071289, "learning_rate": 2.18375e-06, "loss": 0.1601, "step": 1750 }, { "epoch": 0.8123569794050344, "grad_norm": 5.27864408493042, "learning_rate": 2.215e-06, "loss": 0.1656, "step": 1775 }, { "epoch": 0.8237986270022883, "grad_norm": 9.679104804992676, "learning_rate": 2.24625e-06, "loss": 0.1842, "step": 1800 }, { "epoch": 0.8352402745995423, "grad_norm": 4.940381050109863, "learning_rate": 2.2775000000000002e-06, "loss": 0.1455, "step": 1825 }, { "epoch": 0.8466819221967964, "grad_norm": 15.962414741516113, "learning_rate": 2.30875e-06, "loss": 0.1936, "step": 1850 }, { "epoch": 0.8581235697940504, "grad_norm": 6.16251802444458, "learning_rate": 2.34e-06, "loss": 0.1623, "step": 1875 }, { "epoch": 0.8695652173913043, "grad_norm": 16.07195281982422, "learning_rate": 2.3712500000000004e-06, "loss": 0.1932, "step": 1900 }, { "epoch": 0.8810068649885584, "grad_norm": 5.088130950927734, "learning_rate": 2.4025000000000003e-06, "loss": 0.1568, "step": 1925 }, { "epoch": 0.8924485125858124, "grad_norm": 10.709908485412598, "learning_rate": 2.43375e-06, "loss": 0.165, "step": 1950 }, { "epoch": 0.9038901601830663, "grad_norm": 6.499959945678711, "learning_rate": 2.465e-06, "loss": 0.1385, "step": 1975 }, { "epoch": 0.9153318077803204, "grad_norm": 13.85750675201416, "learning_rate": 2.49625e-06, "loss": 0.1911, "step": 2000 }, { "epoch": 0.9267734553775744, "grad_norm": 5.7312822341918945, "learning_rate": 2.5275e-06, "loss": 0.1525, "step": 2025 }, { "epoch": 0.9382151029748284, "grad_norm": 12.252706527709961, "learning_rate": 2.55875e-06, "loss": 0.2005, "step": 2050 }, { "epoch": 0.9496567505720824, "grad_norm": 11.799100875854492, "learning_rate": 2.59e-06, "loss": 0.1695, "step": 2075 }, { "epoch": 0.9610983981693364, "grad_norm": 16.282901763916016, "learning_rate": 2.62125e-06, "loss": 0.2019, "step": 2100 }, { "epoch": 0.9725400457665904, "grad_norm": 4.300443172454834, "learning_rate": 2.6525e-06, "loss": 0.1514, "step": 2125 }, { "epoch": 0.9839816933638444, "grad_norm": 17.060945510864258, "learning_rate": 2.6837500000000004e-06, "loss": 0.1909, "step": 2150 }, { "epoch": 0.9954233409610984, "grad_norm": 8.843390464782715, "learning_rate": 2.7150000000000003e-06, "loss": 0.1605, "step": 2175 }, { "epoch": 1.0068649885583525, "grad_norm": 3.554812431335449, "learning_rate": 2.74625e-06, "loss": 0.1245, "step": 2200 }, { "epoch": 1.0183066361556063, "grad_norm": 5.7890305519104, "learning_rate": 2.7775e-06, "loss": 0.0978, "step": 2225 }, { "epoch": 1.0297482837528604, "grad_norm": 2.891451358795166, "learning_rate": 2.8087500000000004e-06, "loss": 0.1469, "step": 2250 }, { "epoch": 1.0411899313501145, "grad_norm": 5.289301872253418, "learning_rate": 2.8400000000000003e-06, "loss": 0.1022, "step": 2275 }, { "epoch": 1.0526315789473684, "grad_norm": 6.965038299560547, "learning_rate": 2.87125e-06, "loss": 0.134, "step": 2300 }, { "epoch": 1.0640732265446224, "grad_norm": 18.425371170043945, "learning_rate": 2.9025e-06, "loss": 0.0921, "step": 2325 }, { "epoch": 1.0755148741418765, "grad_norm": 4.118972301483154, "learning_rate": 2.93375e-06, "loss": 0.1319, "step": 2350 }, { "epoch": 1.0869565217391304, "grad_norm": 5.5514702796936035, "learning_rate": 2.965e-06, "loss": 0.0917, "step": 2375 }, { "epoch": 1.0983981693363845, "grad_norm": 3.7737677097320557, "learning_rate": 2.99625e-06, "loss": 0.1433, "step": 2400 }, { "epoch": 1.1098398169336385, "grad_norm": 5.6356987953186035, "learning_rate": 3.0275000000000002e-06, "loss": 0.1072, "step": 2425 }, { "epoch": 1.1212814645308924, "grad_norm": 3.293067693710327, "learning_rate": 3.05875e-06, "loss": 0.1306, "step": 2450 }, { "epoch": 1.1327231121281465, "grad_norm": 8.140530586242676, "learning_rate": 3.09e-06, "loss": 0.0957, "step": 2475 }, { "epoch": 1.1441647597254005, "grad_norm": 12.179439544677734, "learning_rate": 3.1212500000000004e-06, "loss": 0.1193, "step": 2500 }, { "epoch": 1.1556064073226544, "grad_norm": 5.836298942565918, "learning_rate": 3.1525e-06, "loss": 0.1076, "step": 2525 }, { "epoch": 1.1670480549199085, "grad_norm": 5.364569664001465, "learning_rate": 3.18375e-06, "loss": 0.1626, "step": 2550 }, { "epoch": 1.1784897025171626, "grad_norm": 5.4423723220825195, "learning_rate": 3.215e-06, "loss": 0.1094, "step": 2575 }, { "epoch": 1.1899313501144164, "grad_norm": 3.647977590560913, "learning_rate": 3.24625e-06, "loss": 0.129, "step": 2600 }, { "epoch": 1.2013729977116705, "grad_norm": 6.781228065490723, "learning_rate": 3.2775e-06, "loss": 0.1072, "step": 2625 }, { "epoch": 1.2128146453089246, "grad_norm": 3.458967924118042, "learning_rate": 3.30875e-06, "loss": 0.1365, "step": 2650 }, { "epoch": 1.2242562929061784, "grad_norm": 7.483636856079102, "learning_rate": 3.34e-06, "loss": 0.1026, "step": 2675 }, { "epoch": 1.2356979405034325, "grad_norm": 5.426140308380127, "learning_rate": 3.37125e-06, "loss": 0.1561, "step": 2700 }, { "epoch": 1.2471395881006866, "grad_norm": 6.9238104820251465, "learning_rate": 3.4025e-06, "loss": 0.1244, "step": 2725 }, { "epoch": 1.2585812356979404, "grad_norm": 3.6151397228240967, "learning_rate": 3.4337500000000004e-06, "loss": 0.1412, "step": 2750 }, { "epoch": 1.2700228832951945, "grad_norm": 8.311691284179688, "learning_rate": 3.4650000000000003e-06, "loss": 0.108, "step": 2775 }, { "epoch": 1.2814645308924484, "grad_norm": 3.9945647716522217, "learning_rate": 3.49625e-06, "loss": 0.1473, "step": 2800 }, { "epoch": 1.2929061784897025, "grad_norm": 8.727825164794922, "learning_rate": 3.5275e-06, "loss": 0.1101, "step": 2825 }, { "epoch": 1.3043478260869565, "grad_norm": 6.813472270965576, "learning_rate": 3.5587500000000004e-06, "loss": 0.1693, "step": 2850 }, { "epoch": 1.3157894736842106, "grad_norm": 7.848366737365723, "learning_rate": 3.5900000000000004e-06, "loss": 0.1067, "step": 2875 }, { "epoch": 1.3272311212814645, "grad_norm": 7.600481986999512, "learning_rate": 3.6212500000000003e-06, "loss": 0.1386, "step": 2900 }, { "epoch": 1.3386727688787186, "grad_norm": 6.826284885406494, "learning_rate": 3.6525e-06, "loss": 0.1102, "step": 2925 }, { "epoch": 1.3501144164759724, "grad_norm": 6.185801029205322, "learning_rate": 3.6837500000000005e-06, "loss": 0.1441, "step": 2950 }, { "epoch": 1.3615560640732265, "grad_norm": 7.203296661376953, "learning_rate": 3.7150000000000004e-06, "loss": 0.1033, "step": 2975 }, { "epoch": 1.3729977116704806, "grad_norm": 4.766386032104492, "learning_rate": 3.7462500000000003e-06, "loss": 0.152, "step": 3000 }, { "epoch": 1.3844393592677346, "grad_norm": 5.796128273010254, "learning_rate": 3.7775000000000007e-06, "loss": 0.0942, "step": 3025 }, { "epoch": 1.3958810068649885, "grad_norm": 3.6194448471069336, "learning_rate": 3.8087500000000006e-06, "loss": 0.1314, "step": 3050 }, { "epoch": 1.4073226544622426, "grad_norm": 1.8742519617080688, "learning_rate": 3.84e-06, "loss": 0.0959, "step": 3075 }, { "epoch": 1.4187643020594964, "grad_norm": 5.452184200286865, "learning_rate": 3.8712499999999996e-06, "loss": 0.1477, "step": 3100 }, { "epoch": 1.4302059496567505, "grad_norm": 9.028647422790527, "learning_rate": 3.9025e-06, "loss": 0.0983, "step": 3125 }, { "epoch": 1.4416475972540046, "grad_norm": 5.759121417999268, "learning_rate": 3.93375e-06, "loss": 0.1264, "step": 3150 }, { "epoch": 1.4530892448512587, "grad_norm": 4.013973236083984, "learning_rate": 3.965e-06, "loss": 0.1103, "step": 3175 }, { "epoch": 1.4645308924485125, "grad_norm": 4.706231594085693, "learning_rate": 3.99625e-06, "loss": 0.1425, "step": 3200 }, { "epoch": 1.4759725400457666, "grad_norm": 7.723937511444092, "learning_rate": 4.0275e-06, "loss": 0.0989, "step": 3225 }, { "epoch": 1.4874141876430205, "grad_norm": 4.614337921142578, "learning_rate": 4.05875e-06, "loss": 0.1456, "step": 3250 }, { "epoch": 1.4988558352402745, "grad_norm": 6.104960918426514, "learning_rate": 4.09e-06, "loss": 0.1257, "step": 3275 }, { "epoch": 1.5102974828375286, "grad_norm": 3.505063772201538, "learning_rate": 4.12125e-06, "loss": 0.1506, "step": 3300 }, { "epoch": 1.5217391304347827, "grad_norm": 11.545547485351562, "learning_rate": 4.1525000000000005e-06, "loss": 0.1014, "step": 3325 }, { "epoch": 1.5331807780320366, "grad_norm": 5.136053085327148, "learning_rate": 4.18375e-06, "loss": 0.1383, "step": 3350 }, { "epoch": 1.5446224256292906, "grad_norm": 7.054224491119385, "learning_rate": 4.215e-06, "loss": 0.1002, "step": 3375 }, { "epoch": 1.5560640732265445, "grad_norm": 6.568813800811768, "learning_rate": 4.24625e-06, "loss": 0.1384, "step": 3400 }, { "epoch": 1.5675057208237986, "grad_norm": 5.4689202308654785, "learning_rate": 4.2775e-06, "loss": 0.1155, "step": 3425 }, { "epoch": 1.5789473684210527, "grad_norm": 3.3778083324432373, "learning_rate": 4.30875e-06, "loss": 0.1466, "step": 3450 }, { "epoch": 1.5903890160183067, "grad_norm": 6.343507766723633, "learning_rate": 4.34e-06, "loss": 0.1067, "step": 3475 }, { "epoch": 1.6018306636155606, "grad_norm": 4.513115882873535, "learning_rate": 4.371250000000001e-06, "loss": 0.1327, "step": 3500 }, { "epoch": 1.6132723112128147, "grad_norm": 7.246913909912109, "learning_rate": 4.402500000000001e-06, "loss": 0.1074, "step": 3525 }, { "epoch": 1.6247139588100685, "grad_norm": 6.542562961578369, "learning_rate": 4.4337500000000005e-06, "loss": 0.1428, "step": 3550 }, { "epoch": 1.6361556064073226, "grad_norm": 5.578348159790039, "learning_rate": 4.4650000000000004e-06, "loss": 0.1148, "step": 3575 }, { "epoch": 1.6475972540045767, "grad_norm": 5.8001885414123535, "learning_rate": 4.49625e-06, "loss": 0.1406, "step": 3600 }, { "epoch": 1.6590389016018308, "grad_norm": 9.58809757232666, "learning_rate": 4.5275e-06, "loss": 0.1139, "step": 3625 }, { "epoch": 1.6704805491990846, "grad_norm": 5.428323268890381, "learning_rate": 4.55875e-06, "loss": 0.1471, "step": 3650 }, { "epoch": 1.6819221967963387, "grad_norm": 8.763864517211914, "learning_rate": 4.590000000000001e-06, "loss": 0.1075, "step": 3675 }, { "epoch": 1.6933638443935926, "grad_norm": 3.4743332862854004, "learning_rate": 4.62125e-06, "loss": 0.1537, "step": 3700 }, { "epoch": 1.7048054919908466, "grad_norm": 6.93314790725708, "learning_rate": 4.6525e-06, "loss": 0.1257, "step": 3725 }, { "epoch": 1.7162471395881007, "grad_norm": 3.2530903816223145, "learning_rate": 4.68375e-06, "loss": 0.1429, "step": 3750 }, { "epoch": 1.7276887871853548, "grad_norm": 5.646566390991211, "learning_rate": 4.715e-06, "loss": 0.1167, "step": 3775 }, { "epoch": 1.7391304347826086, "grad_norm": 4.806797981262207, "learning_rate": 4.74625e-06, "loss": 0.1387, "step": 3800 }, { "epoch": 1.7505720823798627, "grad_norm": 8.664536476135254, "learning_rate": 4.7775e-06, "loss": 0.1173, "step": 3825 }, { "epoch": 1.7620137299771166, "grad_norm": 2.8637678623199463, "learning_rate": 4.80875e-06, "loss": 0.1539, "step": 3850 }, { "epoch": 1.7734553775743707, "grad_norm": 7.306221961975098, "learning_rate": 4.84e-06, "loss": 0.1342, "step": 3875 }, { "epoch": 1.7848970251716247, "grad_norm": 3.557720184326172, "learning_rate": 4.87125e-06, "loss": 0.1611, "step": 3900 }, { "epoch": 1.7963386727688788, "grad_norm": 10.298198699951172, "learning_rate": 4.9025e-06, "loss": 0.1093, "step": 3925 }, { "epoch": 1.8077803203661327, "grad_norm": 3.6880478858947754, "learning_rate": 4.93375e-06, "loss": 0.1332, "step": 3950 }, { "epoch": 1.8192219679633868, "grad_norm": 5.447417259216309, "learning_rate": 4.965e-06, "loss": 0.1232, "step": 3975 }, { "epoch": 1.8306636155606406, "grad_norm": 7.719545364379883, "learning_rate": 4.996250000000001e-06, "loss": 0.1445, "step": 4000 }, { "epoch": 1.8421052631578947, "grad_norm": 5.616191387176514, "learning_rate": 5.0275000000000006e-06, "loss": 0.1335, "step": 4025 }, { "epoch": 1.8535469107551488, "grad_norm": 4.927710056304932, "learning_rate": 5.0587500000000005e-06, "loss": 0.139, "step": 4050 }, { "epoch": 1.8649885583524028, "grad_norm": 9.265386581420898, "learning_rate": 5.09e-06, "loss": 0.1206, "step": 4075 }, { "epoch": 1.8764302059496567, "grad_norm": 6.632187366485596, "learning_rate": 5.12125e-06, "loss": 0.1312, "step": 4100 }, { "epoch": 1.8878718535469108, "grad_norm": 11.779533386230469, "learning_rate": 5.1525e-06, "loss": 0.1116, "step": 4125 }, { "epoch": 1.8993135011441646, "grad_norm": 3.810981035232544, "learning_rate": 5.182500000000001e-06, "loss": 0.1401, "step": 4150 }, { "epoch": 1.9107551487414187, "grad_norm": 4.736062526702881, "learning_rate": 5.213750000000001e-06, "loss": 0.1171, "step": 4175 }, { "epoch": 1.9221967963386728, "grad_norm": 5.427768707275391, "learning_rate": 5.245e-06, "loss": 0.1392, "step": 4200 }, { "epoch": 1.9336384439359269, "grad_norm": 10.339317321777344, "learning_rate": 5.27625e-06, "loss": 0.1249, "step": 4225 }, { "epoch": 1.9450800915331807, "grad_norm": 5.156798362731934, "learning_rate": 5.3075e-06, "loss": 0.1413, "step": 4250 }, { "epoch": 1.9565217391304348, "grad_norm": 6.890321731567383, "learning_rate": 5.33875e-06, "loss": 0.1059, "step": 4275 }, { "epoch": 1.9679633867276887, "grad_norm": 4.059100151062012, "learning_rate": 5.37e-06, "loss": 0.1443, "step": 4300 }, { "epoch": 1.9794050343249427, "grad_norm": 12.56505012512207, "learning_rate": 5.40125e-06, "loss": 0.1082, "step": 4325 }, { "epoch": 1.9908466819221968, "grad_norm": 5.877668857574463, "learning_rate": 5.4325e-06, "loss": 0.1328, "step": 4350 }, { "epoch": 2.002288329519451, "grad_norm": 3.3472976684570312, "learning_rate": 5.46375e-06, "loss": 0.1337, "step": 4375 }, { "epoch": 2.013729977116705, "grad_norm": 1.5074313879013062, "learning_rate": 5.495e-06, "loss": 0.0594, "step": 4400 }, { "epoch": 2.0251716247139586, "grad_norm": 4.803617000579834, "learning_rate": 5.52625e-06, "loss": 0.0803, "step": 4425 }, { "epoch": 2.0366132723112127, "grad_norm": 3.4739410877227783, "learning_rate": 5.557500000000001e-06, "loss": 0.0581, "step": 4450 }, { "epoch": 2.0480549199084668, "grad_norm": 2.3615224361419678, "learning_rate": 5.5887500000000005e-06, "loss": 0.0815, "step": 4475 }, { "epoch": 2.059496567505721, "grad_norm": 2.8116395473480225, "learning_rate": 5.62e-06, "loss": 0.066, "step": 4500 }, { "epoch": 2.070938215102975, "grad_norm": 3.5427844524383545, "learning_rate": 5.65125e-06, "loss": 0.0756, "step": 4525 }, { "epoch": 2.082379862700229, "grad_norm": 10.534377098083496, "learning_rate": 5.6825e-06, "loss": 0.0608, "step": 4550 }, { "epoch": 2.0938215102974826, "grad_norm": 3.33672833442688, "learning_rate": 5.71375e-06, "loss": 0.0937, "step": 4575 }, { "epoch": 2.1052631578947367, "grad_norm": 3.8001768589019775, "learning_rate": 5.745e-06, "loss": 0.058, "step": 4600 }, { "epoch": 2.116704805491991, "grad_norm": 4.036559104919434, "learning_rate": 5.776250000000001e-06, "loss": 0.0789, "step": 4625 }, { "epoch": 2.128146453089245, "grad_norm": 2.8355629444122314, "learning_rate": 5.807500000000001e-06, "loss": 0.0503, "step": 4650 }, { "epoch": 2.139588100686499, "grad_norm": 6.225577354431152, "learning_rate": 5.838750000000001e-06, "loss": 0.0859, "step": 4675 }, { "epoch": 2.151029748283753, "grad_norm": 3.4654979705810547, "learning_rate": 5.8700000000000005e-06, "loss": 0.0635, "step": 4700 }, { "epoch": 2.1624713958810067, "grad_norm": 3.9897263050079346, "learning_rate": 5.9012500000000005e-06, "loss": 0.078, "step": 4725 }, { "epoch": 2.1739130434782608, "grad_norm": 3.4768152236938477, "learning_rate": 5.9325e-06, "loss": 0.0707, "step": 4750 }, { "epoch": 2.185354691075515, "grad_norm": 4.6715192794799805, "learning_rate": 5.96375e-06, "loss": 0.0937, "step": 4775 }, { "epoch": 2.196796338672769, "grad_norm": 3.6018433570861816, "learning_rate": 5.995e-06, "loss": 0.0616, "step": 4800 }, { "epoch": 2.208237986270023, "grad_norm": 4.695530891418457, "learning_rate": 6.02625e-06, "loss": 0.0853, "step": 4825 }, { "epoch": 2.219679633867277, "grad_norm": 1.7175928354263306, "learning_rate": 6.0575e-06, "loss": 0.0527, "step": 4850 }, { "epoch": 2.2311212814645307, "grad_norm": 2.272045850753784, "learning_rate": 6.08875e-06, "loss": 0.1014, "step": 4875 }, { "epoch": 2.242562929061785, "grad_norm": 9.221810340881348, "learning_rate": 6.12e-06, "loss": 0.0698, "step": 4900 }, { "epoch": 2.254004576659039, "grad_norm": 3.7419273853302, "learning_rate": 6.15125e-06, "loss": 0.1009, "step": 4925 }, { "epoch": 2.265446224256293, "grad_norm": 4.7529802322387695, "learning_rate": 6.1825e-06, "loss": 0.0715, "step": 4950 }, { "epoch": 2.276887871853547, "grad_norm": 3.084136486053467, "learning_rate": 6.2137500000000004e-06, "loss": 0.1042, "step": 4975 }, { "epoch": 2.288329519450801, "grad_norm": 6.452476501464844, "learning_rate": 6.245e-06, "loss": 0.0697, "step": 5000 }, { "epoch": 2.288329519450801, "eval_loss": 0.16196583211421967, "eval_runtime": 5941.1089, "eval_samples_per_second": 1.603, "eval_steps_per_second": 0.2, "eval_wer": 0.10273884942727331, "step": 5000 }, { "epoch": 2.2997711670480547, "grad_norm": 3.5335402488708496, "learning_rate": 6.2434375e-06, "loss": 0.0812, "step": 5025 }, { "epoch": 2.311212814645309, "grad_norm": 7.161555290222168, "learning_rate": 6.235625e-06, "loss": 0.0688, "step": 5050 }, { "epoch": 2.322654462242563, "grad_norm": 2.389561653137207, "learning_rate": 6.2278125e-06, "loss": 0.0872, "step": 5075 }, { "epoch": 2.334096109839817, "grad_norm": 6.461544513702393, "learning_rate": 6.22e-06, "loss": 0.0598, "step": 5100 }, { "epoch": 2.345537757437071, "grad_norm": 4.625803470611572, "learning_rate": 6.2121875e-06, "loss": 0.0825, "step": 5125 }, { "epoch": 2.356979405034325, "grad_norm": 3.9252727031707764, "learning_rate": 6.204375e-06, "loss": 0.0631, "step": 5150 }, { "epoch": 2.3684210526315788, "grad_norm": 2.284882068634033, "learning_rate": 6.196562500000001e-06, "loss": 0.085, "step": 5175 }, { "epoch": 2.379862700228833, "grad_norm": 5.8602399826049805, "learning_rate": 6.18875e-06, "loss": 0.0697, "step": 5200 }, { "epoch": 2.391304347826087, "grad_norm": 5.568091869354248, "learning_rate": 6.1809375000000005e-06, "loss": 0.0872, "step": 5225 }, { "epoch": 2.402745995423341, "grad_norm": 7.008312225341797, "learning_rate": 6.173125e-06, "loss": 0.0661, "step": 5250 }, { "epoch": 2.414187643020595, "grad_norm": 3.3845436573028564, "learning_rate": 6.165312500000001e-06, "loss": 0.0912, "step": 5275 }, { "epoch": 2.425629290617849, "grad_norm": 3.0679447650909424, "learning_rate": 6.1575e-06, "loss": 0.0723, "step": 5300 }, { "epoch": 2.437070938215103, "grad_norm": 5.007796287536621, "learning_rate": 6.1496875000000006e-06, "loss": 0.0937, "step": 5325 }, { "epoch": 2.448512585812357, "grad_norm": 4.2327046394348145, "learning_rate": 6.141875e-06, "loss": 0.073, "step": 5350 }, { "epoch": 2.459954233409611, "grad_norm": 4.146753311157227, "learning_rate": 6.1340625e-06, "loss": 0.0894, "step": 5375 }, { "epoch": 2.471395881006865, "grad_norm": 1.6775671243667603, "learning_rate": 6.12625e-06, "loss": 0.0866, "step": 5400 }, { "epoch": 2.482837528604119, "grad_norm": 3.461623191833496, "learning_rate": 6.1184375e-06, "loss": 0.0859, "step": 5425 }, { "epoch": 2.494279176201373, "grad_norm": 3.211845636367798, "learning_rate": 6.1106250000000005e-06, "loss": 0.0732, "step": 5450 }, { "epoch": 2.505720823798627, "grad_norm": 4.490880489349365, "learning_rate": 6.1028125e-06, "loss": 0.0871, "step": 5475 }, { "epoch": 2.517162471395881, "grad_norm": 4.470268249511719, "learning_rate": 6.095e-06, "loss": 0.0734, "step": 5500 }, { "epoch": 2.528604118993135, "grad_norm": 5.099210262298584, "learning_rate": 6.0871875e-06, "loss": 0.0961, "step": 5525 }, { "epoch": 2.540045766590389, "grad_norm": 2.920400381088257, "learning_rate": 6.0793750000000006e-06, "loss": 0.0712, "step": 5550 }, { "epoch": 2.551487414187643, "grad_norm": 4.1212592124938965, "learning_rate": 6.0715625e-06, "loss": 0.0911, "step": 5575 }, { "epoch": 2.5629290617848968, "grad_norm": 1.774526596069336, "learning_rate": 6.06375e-06, "loss": 0.0664, "step": 5600 }, { "epoch": 2.5743707093821513, "grad_norm": 3.186657428741455, "learning_rate": 6.0559375e-06, "loss": 0.0807, "step": 5625 }, { "epoch": 2.585812356979405, "grad_norm": 1.9952179193496704, "learning_rate": 6.048125000000001e-06, "loss": 0.0776, "step": 5650 }, { "epoch": 2.597254004576659, "grad_norm": 2.9018378257751465, "learning_rate": 6.0403125000000005e-06, "loss": 0.092, "step": 5675 }, { "epoch": 2.608695652173913, "grad_norm": 3.827449083328247, "learning_rate": 6.0325e-06, "loss": 0.0645, "step": 5700 }, { "epoch": 2.620137299771167, "grad_norm": 3.021289825439453, "learning_rate": 6.0246875e-06, "loss": 0.0791, "step": 5725 }, { "epoch": 2.6315789473684212, "grad_norm": 3.3749895095825195, "learning_rate": 6.016875e-06, "loss": 0.075, "step": 5750 }, { "epoch": 2.643020594965675, "grad_norm": 4.381983757019043, "learning_rate": 6.0090625000000005e-06, "loss": 0.0977, "step": 5775 }, { "epoch": 2.654462242562929, "grad_norm": 8.246472358703613, "learning_rate": 6.00125e-06, "loss": 0.0669, "step": 5800 }, { "epoch": 2.665903890160183, "grad_norm": 2.754366874694824, "learning_rate": 5.9934375e-06, "loss": 0.0845, "step": 5825 }, { "epoch": 2.677345537757437, "grad_norm": 3.511594295501709, "learning_rate": 5.985625e-06, "loss": 0.0663, "step": 5850 }, { "epoch": 2.688787185354691, "grad_norm": 3.16924786567688, "learning_rate": 5.977812500000001e-06, "loss": 0.0868, "step": 5875 }, { "epoch": 2.700228832951945, "grad_norm": 3.8349554538726807, "learning_rate": 5.9700000000000004e-06, "loss": 0.0644, "step": 5900 }, { "epoch": 2.7116704805491993, "grad_norm": 2.3711776733398438, "learning_rate": 5.9621875e-06, "loss": 0.0786, "step": 5925 }, { "epoch": 2.723112128146453, "grad_norm": 2.4516854286193848, "learning_rate": 5.954375e-06, "loss": 0.0774, "step": 5950 }, { "epoch": 2.734553775743707, "grad_norm": 2.746248483657837, "learning_rate": 5.946562500000001e-06, "loss": 0.1053, "step": 5975 }, { "epoch": 2.745995423340961, "grad_norm": 5.512765407562256, "learning_rate": 5.9387500000000005e-06, "loss": 0.0624, "step": 6000 }, { "epoch": 2.757437070938215, "grad_norm": 3.6978447437286377, "learning_rate": 5.9309375e-06, "loss": 0.0832, "step": 6025 }, { "epoch": 2.7688787185354693, "grad_norm": 3.7485339641571045, "learning_rate": 5.923125e-06, "loss": 0.0646, "step": 6050 }, { "epoch": 2.780320366132723, "grad_norm": 2.9584758281707764, "learning_rate": 5.9153125e-06, "loss": 0.0797, "step": 6075 }, { "epoch": 2.791762013729977, "grad_norm": 3.063089609146118, "learning_rate": 5.907500000000001e-06, "loss": 0.0594, "step": 6100 }, { "epoch": 2.803203661327231, "grad_norm": 3.4016189575195312, "learning_rate": 5.8996875000000004e-06, "loss": 0.0841, "step": 6125 }, { "epoch": 2.814645308924485, "grad_norm": 5.774538993835449, "learning_rate": 5.8921875e-06, "loss": 0.0851, "step": 6150 }, { "epoch": 2.8260869565217392, "grad_norm": 2.9045443534851074, "learning_rate": 5.884375e-06, "loss": 0.0849, "step": 6175 }, { "epoch": 2.837528604118993, "grad_norm": 9.496201515197754, "learning_rate": 5.8765625000000005e-06, "loss": 0.08, "step": 6200 }, { "epoch": 2.8489702517162474, "grad_norm": 4.791667938232422, "learning_rate": 5.86875e-06, "loss": 0.0912, "step": 6225 }, { "epoch": 2.860411899313501, "grad_norm": 3.803701400756836, "learning_rate": 5.8609375e-06, "loss": 0.0662, "step": 6250 }, { "epoch": 2.871853546910755, "grad_norm": 7.12540864944458, "learning_rate": 5.853125e-06, "loss": 0.0981, "step": 6275 }, { "epoch": 2.883295194508009, "grad_norm": 6.205677032470703, "learning_rate": 5.845312500000001e-06, "loss": 0.0758, "step": 6300 }, { "epoch": 2.8947368421052633, "grad_norm": 4.83319616317749, "learning_rate": 5.8375000000000004e-06, "loss": 0.0805, "step": 6325 }, { "epoch": 2.9061784897025174, "grad_norm": 6.701275825500488, "learning_rate": 5.8296875e-06, "loss": 0.0796, "step": 6350 }, { "epoch": 2.917620137299771, "grad_norm": 4.209991455078125, "learning_rate": 5.821875e-06, "loss": 0.0876, "step": 6375 }, { "epoch": 2.929061784897025, "grad_norm": 3.636922597885132, "learning_rate": 5.814062500000001e-06, "loss": 0.0719, "step": 6400 }, { "epoch": 2.940503432494279, "grad_norm": 4.24993896484375, "learning_rate": 5.8062500000000005e-06, "loss": 0.0848, "step": 6425 }, { "epoch": 2.9519450800915332, "grad_norm": 4.109525203704834, "learning_rate": 5.7984375e-06, "loss": 0.0775, "step": 6450 }, { "epoch": 2.9633867276887873, "grad_norm": 3.515380382537842, "learning_rate": 5.790625e-06, "loss": 0.0949, "step": 6475 }, { "epoch": 2.974828375286041, "grad_norm": 2.294419765472412, "learning_rate": 5.782812500000001e-06, "loss": 0.0721, "step": 6500 }, { "epoch": 2.9862700228832955, "grad_norm": 3.1478168964385986, "learning_rate": 5.775000000000001e-06, "loss": 0.0784, "step": 6525 }, { "epoch": 2.997711670480549, "grad_norm": 6.680340766906738, "learning_rate": 5.7671875e-06, "loss": 0.0817, "step": 6550 }, { "epoch": 3.009153318077803, "grad_norm": 5.719060897827148, "learning_rate": 5.759375e-06, "loss": 0.0532, "step": 6575 }, { "epoch": 3.0205949656750573, "grad_norm": 12.910862922668457, "learning_rate": 5.7515625e-06, "loss": 0.039, "step": 6600 }, { "epoch": 3.0320366132723113, "grad_norm": 2.577425003051758, "learning_rate": 5.743750000000001e-06, "loss": 0.0385, "step": 6625 }, { "epoch": 3.0434782608695654, "grad_norm": 6.705196380615234, "learning_rate": 5.7359375e-06, "loss": 0.0373, "step": 6650 }, { "epoch": 3.054919908466819, "grad_norm": 3.4962539672851562, "learning_rate": 5.728125e-06, "loss": 0.0542, "step": 6675 }, { "epoch": 3.066361556064073, "grad_norm": 7.721837043762207, "learning_rate": 5.7203125e-06, "loss": 0.0384, "step": 6700 }, { "epoch": 3.077803203661327, "grad_norm": 2.549494504928589, "learning_rate": 5.712500000000001e-06, "loss": 0.0443, "step": 6725 }, { "epoch": 3.0892448512585813, "grad_norm": 7.875229358673096, "learning_rate": 5.7046875e-06, "loss": 0.0546, "step": 6750 }, { "epoch": 3.1006864988558354, "grad_norm": 6.519637584686279, "learning_rate": 5.696875e-06, "loss": 0.0572, "step": 6775 }, { "epoch": 3.1121281464530894, "grad_norm": 3.3160250186920166, "learning_rate": 5.6890625e-06, "loss": 0.0268, "step": 6800 }, { "epoch": 3.123569794050343, "grad_norm": 1.0826241970062256, "learning_rate": 5.681250000000001e-06, "loss": 0.0455, "step": 6825 }, { "epoch": 3.135011441647597, "grad_norm": 2.4593420028686523, "learning_rate": 5.6734375e-06, "loss": 0.0374, "step": 6850 }, { "epoch": 3.1464530892448512, "grad_norm": 5.654216289520264, "learning_rate": 5.6656250000000005e-06, "loss": 0.0528, "step": 6875 }, { "epoch": 3.1578947368421053, "grad_norm": 4.648204803466797, "learning_rate": 5.6578125e-06, "loss": 0.0345, "step": 6900 }, { "epoch": 3.1693363844393594, "grad_norm": 4.878942489624023, "learning_rate": 5.65e-06, "loss": 0.0397, "step": 6925 }, { "epoch": 3.1807780320366135, "grad_norm": 5.794234752655029, "learning_rate": 5.642187500000001e-06, "loss": 0.0457, "step": 6950 }, { "epoch": 3.192219679633867, "grad_norm": 3.9493894577026367, "learning_rate": 5.634375e-06, "loss": 0.0547, "step": 6975 }, { "epoch": 3.203661327231121, "grad_norm": 4.193353652954102, "learning_rate": 5.6265625e-06, "loss": 0.0383, "step": 7000 }, { "epoch": 3.2151029748283753, "grad_norm": 3.423971176147461, "learning_rate": 5.61875e-06, "loss": 0.0462, "step": 7025 }, { "epoch": 3.2265446224256293, "grad_norm": 6.13396692276001, "learning_rate": 5.610937500000001e-06, "loss": 0.0473, "step": 7050 }, { "epoch": 3.2379862700228834, "grad_norm": 4.359652996063232, "learning_rate": 5.603125e-06, "loss": 0.0511, "step": 7075 }, { "epoch": 3.2494279176201375, "grad_norm": 1.4082869291305542, "learning_rate": 5.5953125000000005e-06, "loss": 0.0423, "step": 7100 }, { "epoch": 3.260869565217391, "grad_norm": 4.264839172363281, "learning_rate": 5.5875e-06, "loss": 0.0462, "step": 7125 }, { "epoch": 3.272311212814645, "grad_norm": 7.262903213500977, "learning_rate": 5.579687500000001e-06, "loss": 0.0478, "step": 7150 }, { "epoch": 3.2837528604118993, "grad_norm": 3.770082950592041, "learning_rate": 5.571875e-06, "loss": 0.0361, "step": 7175 }, { "epoch": 3.2951945080091534, "grad_norm": 7.296878814697266, "learning_rate": 5.5640625000000006e-06, "loss": 0.0432, "step": 7200 }, { "epoch": 3.3066361556064074, "grad_norm": 2.4708614349365234, "learning_rate": 5.55625e-06, "loss": 0.0599, "step": 7225 }, { "epoch": 3.3180778032036615, "grad_norm": 4.431317329406738, "learning_rate": 5.5484375e-06, "loss": 0.0424, "step": 7250 }, { "epoch": 3.329519450800915, "grad_norm": 4.9188127517700195, "learning_rate": 5.540625e-06, "loss": 0.0514, "step": 7275 }, { "epoch": 3.3409610983981692, "grad_norm": 5.590330123901367, "learning_rate": 5.5328125e-06, "loss": 0.0502, "step": 7300 }, { "epoch": 3.3524027459954233, "grad_norm": 1.5474858283996582, "learning_rate": 5.5250000000000005e-06, "loss": 0.0506, "step": 7325 }, { "epoch": 3.3638443935926774, "grad_norm": 6.954276084899902, "learning_rate": 5.5171875e-06, "loss": 0.0432, "step": 7350 }, { "epoch": 3.3752860411899315, "grad_norm": 2.652070999145508, "learning_rate": 5.509375e-06, "loss": 0.0516, "step": 7375 }, { "epoch": 3.386727688787185, "grad_norm": 1.9480232000350952, "learning_rate": 5.5015625e-06, "loss": 0.0408, "step": 7400 }, { "epoch": 3.398169336384439, "grad_norm": 1.5268117189407349, "learning_rate": 5.4937500000000006e-06, "loss": 0.0468, "step": 7425 }, { "epoch": 3.4096109839816933, "grad_norm": 3.8624870777130127, "learning_rate": 5.4859375e-06, "loss": 0.0401, "step": 7450 }, { "epoch": 3.4210526315789473, "grad_norm": 3.136197805404663, "learning_rate": 5.478125e-06, "loss": 0.0485, "step": 7475 }, { "epoch": 3.4324942791762014, "grad_norm": 4.95338249206543, "learning_rate": 5.4703125e-06, "loss": 0.0381, "step": 7500 }, { "epoch": 3.4439359267734555, "grad_norm": 4.111715793609619, "learning_rate": 5.462500000000001e-06, "loss": 0.0484, "step": 7525 }, { "epoch": 3.4553775743707096, "grad_norm": 8.481527328491211, "learning_rate": 5.4546875000000004e-06, "loss": 0.0573, "step": 7550 }, { "epoch": 3.466819221967963, "grad_norm": 2.4452638626098633, "learning_rate": 5.446875e-06, "loss": 0.0511, "step": 7575 }, { "epoch": 3.4782608695652173, "grad_norm": 4.839463710784912, "learning_rate": 5.4390625e-06, "loss": 0.0454, "step": 7600 }, { "epoch": 3.4897025171624714, "grad_norm": 3.3421833515167236, "learning_rate": 5.43125e-06, "loss": 0.0531, "step": 7625 }, { "epoch": 3.5011441647597255, "grad_norm": 5.8438334465026855, "learning_rate": 5.4234375000000005e-06, "loss": 0.0464, "step": 7650 }, { "epoch": 3.5125858123569795, "grad_norm": 3.1992125511169434, "learning_rate": 5.415625e-06, "loss": 0.0558, "step": 7675 }, { "epoch": 3.524027459954233, "grad_norm": 2.4741952419281006, "learning_rate": 5.4078125e-06, "loss": 0.0496, "step": 7700 }, { "epoch": 3.5354691075514877, "grad_norm": 2.3519113063812256, "learning_rate": 5.4e-06, "loss": 0.0517, "step": 7725 }, { "epoch": 3.5469107551487413, "grad_norm": 3.754331111907959, "learning_rate": 5.392187500000001e-06, "loss": 0.0417, "step": 7750 }, { "epoch": 3.5583524027459954, "grad_norm": 2.477574586868286, "learning_rate": 5.3843750000000004e-06, "loss": 0.048, "step": 7775 }, { "epoch": 3.5697940503432495, "grad_norm": 4.506994724273682, "learning_rate": 5.3765625e-06, "loss": 0.0393, "step": 7800 }, { "epoch": 3.5812356979405036, "grad_norm": 2.5902769565582275, "learning_rate": 5.36875e-06, "loss": 0.0493, "step": 7825 }, { "epoch": 3.5926773455377576, "grad_norm": 9.635528564453125, "learning_rate": 5.360937500000001e-06, "loss": 0.0456, "step": 7850 }, { "epoch": 3.6041189931350113, "grad_norm": 5.572010040283203, "learning_rate": 5.3531250000000005e-06, "loss": 0.0451, "step": 7875 }, { "epoch": 3.6155606407322654, "grad_norm": 8.494470596313477, "learning_rate": 5.3453125e-06, "loss": 0.0388, "step": 7900 }, { "epoch": 3.6270022883295194, "grad_norm": 3.9414212703704834, "learning_rate": 5.3375e-06, "loss": 0.0555, "step": 7925 }, { "epoch": 3.6384439359267735, "grad_norm": 4.169886112213135, "learning_rate": 5.3296875e-06, "loss": 0.0468, "step": 7950 }, { "epoch": 3.6498855835240276, "grad_norm": 3.8904569149017334, "learning_rate": 5.321875000000001e-06, "loss": 0.0566, "step": 7975 }, { "epoch": 3.6613272311212812, "grad_norm": 4.280102729797363, "learning_rate": 5.3140624999999996e-06, "loss": 0.0416, "step": 8000 }, { "epoch": 3.6727688787185357, "grad_norm": 3.302794933319092, "learning_rate": 5.30625e-06, "loss": 0.0451, "step": 8025 }, { "epoch": 3.6842105263157894, "grad_norm": 9.184480667114258, "learning_rate": 5.2984375e-06, "loss": 0.0502, "step": 8050 }, { "epoch": 3.6956521739130435, "grad_norm": 4.913455009460449, "learning_rate": 5.290625000000001e-06, "loss": 0.0452, "step": 8075 }, { "epoch": 3.7070938215102975, "grad_norm": 3.561953544616699, "learning_rate": 5.2828125e-06, "loss": 0.0485, "step": 8100 }, { "epoch": 3.7185354691075516, "grad_norm": 3.3016109466552734, "learning_rate": 5.275e-06, "loss": 0.0471, "step": 8125 }, { "epoch": 3.7299771167048057, "grad_norm": 4.43864107131958, "learning_rate": 5.2671875e-06, "loss": 0.0499, "step": 8150 }, { "epoch": 3.7414187643020593, "grad_norm": 6.333988666534424, "learning_rate": 5.259687500000001e-06, "loss": 0.0554, "step": 8175 }, { "epoch": 3.7528604118993134, "grad_norm": 11.240910530090332, "learning_rate": 5.2518750000000004e-06, "loss": 0.034, "step": 8200 }, { "epoch": 3.7643020594965675, "grad_norm": 3.428675889968872, "learning_rate": 5.2440625e-06, "loss": 0.0474, "step": 8225 }, { "epoch": 3.7757437070938216, "grad_norm": 2.0469939708709717, "learning_rate": 5.23625e-06, "loss": 0.0366, "step": 8250 }, { "epoch": 3.7871853546910756, "grad_norm": 2.763183116912842, "learning_rate": 5.228437500000001e-06, "loss": 0.0451, "step": 8275 }, { "epoch": 3.7986270022883293, "grad_norm": 5.365248203277588, "learning_rate": 5.2206250000000005e-06, "loss": 0.0409, "step": 8300 }, { "epoch": 3.8100686498855834, "grad_norm": 3.84425687789917, "learning_rate": 5.2128125e-06, "loss": 0.0493, "step": 8325 }, { "epoch": 3.8215102974828374, "grad_norm": 8.628647804260254, "learning_rate": 5.205e-06, "loss": 0.0468, "step": 8350 }, { "epoch": 3.8329519450800915, "grad_norm": 6.355470657348633, "learning_rate": 5.1971875e-06, "loss": 0.0532, "step": 8375 }, { "epoch": 3.8443935926773456, "grad_norm": 5.792243003845215, "learning_rate": 5.189375000000001e-06, "loss": 0.0389, "step": 8400 }, { "epoch": 3.8558352402745997, "grad_norm": 5.9728312492370605, "learning_rate": 5.1815624999999996e-06, "loss": 0.0495, "step": 8425 }, { "epoch": 3.8672768878718538, "grad_norm": 3.019531488418579, "learning_rate": 5.17375e-06, "loss": 0.0474, "step": 8450 }, { "epoch": 3.8787185354691074, "grad_norm": 4.252594947814941, "learning_rate": 5.1659375e-06, "loss": 0.0571, "step": 8475 }, { "epoch": 3.8901601830663615, "grad_norm": 2.1574008464813232, "learning_rate": 5.158125000000001e-06, "loss": 0.0353, "step": 8500 }, { "epoch": 3.9016018306636155, "grad_norm": 1.5579230785369873, "learning_rate": 5.1503125e-06, "loss": 0.04, "step": 8525 }, { "epoch": 3.9130434782608696, "grad_norm": 9.67751693725586, "learning_rate": 5.1425e-06, "loss": 0.0446, "step": 8550 }, { "epoch": 3.9244851258581237, "grad_norm": 3.2914412021636963, "learning_rate": 5.1346875e-06, "loss": 0.0491, "step": 8575 }, { "epoch": 3.9359267734553773, "grad_norm": 5.403482437133789, "learning_rate": 5.126875000000001e-06, "loss": 0.0501, "step": 8600 }, { "epoch": 3.9473684210526314, "grad_norm": 4.028888702392578, "learning_rate": 5.1190625e-06, "loss": 0.0554, "step": 8625 }, { "epoch": 3.9588100686498855, "grad_norm": 4.119118690490723, "learning_rate": 5.11125e-06, "loss": 0.0411, "step": 8650 }, { "epoch": 3.9702517162471396, "grad_norm": 3.5604629516601562, "learning_rate": 5.1034375e-06, "loss": 0.0595, "step": 8675 }, { "epoch": 3.9816933638443937, "grad_norm": 4.4073686599731445, "learning_rate": 5.095625e-06, "loss": 0.0362, "step": 8700 }, { "epoch": 3.9931350114416477, "grad_norm": 3.6186680793762207, "learning_rate": 5.087812500000001e-06, "loss": 0.05, "step": 8725 }, { "epoch": 4.004576659038902, "grad_norm": 2.3267414569854736, "learning_rate": 5.08e-06, "loss": 0.0403, "step": 8750 }, { "epoch": 4.016018306636155, "grad_norm": 3.6123080253601074, "learning_rate": 5.0721875e-06, "loss": 0.0205, "step": 8775 }, { "epoch": 4.02745995423341, "grad_norm": 4.297911643981934, "learning_rate": 5.064375e-06, "loss": 0.0269, "step": 8800 }, { "epoch": 4.038901601830664, "grad_norm": 3.034499168395996, "learning_rate": 5.056562500000001e-06, "loss": 0.0195, "step": 8825 }, { "epoch": 4.050343249427917, "grad_norm": 2.369044780731201, "learning_rate": 5.04875e-06, "loss": 0.0293, "step": 8850 }, { "epoch": 4.061784897025172, "grad_norm": 2.833590269088745, "learning_rate": 5.0409375e-06, "loss": 0.0187, "step": 8875 }, { "epoch": 4.073226544622425, "grad_norm": 2.4066593647003174, "learning_rate": 5.033125e-06, "loss": 0.0298, "step": 8900 }, { "epoch": 4.08466819221968, "grad_norm": 2.521658420562744, "learning_rate": 5.025312500000001e-06, "loss": 0.0219, "step": 8925 }, { "epoch": 4.0961098398169336, "grad_norm": 2.493262529373169, "learning_rate": 5.0175e-06, "loss": 0.0319, "step": 8950 }, { "epoch": 4.107551487414188, "grad_norm": 4.92982816696167, "learning_rate": 5.0096875000000005e-06, "loss": 0.0229, "step": 8975 }, { "epoch": 4.118993135011442, "grad_norm": 5.053268909454346, "learning_rate": 5.001875e-06, "loss": 0.0348, "step": 9000 }, { "epoch": 4.130434782608695, "grad_norm": 1.4297044277191162, "learning_rate": 4.9940625e-06, "loss": 0.0204, "step": 9025 }, { "epoch": 4.14187643020595, "grad_norm": 1.9617934226989746, "learning_rate": 4.98625e-06, "loss": 0.0369, "step": 9050 }, { "epoch": 4.1533180778032035, "grad_norm": 3.376234531402588, "learning_rate": 4.9784375e-06, "loss": 0.0286, "step": 9075 }, { "epoch": 4.164759725400458, "grad_norm": 1.698594570159912, "learning_rate": 4.970625e-06, "loss": 0.035, "step": 9100 }, { "epoch": 4.176201372997712, "grad_norm": 1.5813435316085815, "learning_rate": 4.9628125e-06, "loss": 0.0215, "step": 9125 }, { "epoch": 4.187643020594965, "grad_norm": 1.6306517124176025, "learning_rate": 4.955e-06, "loss": 0.0401, "step": 9150 }, { "epoch": 4.19908466819222, "grad_norm": 2.4657962322235107, "learning_rate": 4.9471875e-06, "loss": 0.0211, "step": 9175 }, { "epoch": 4.2105263157894735, "grad_norm": 0.9647684693336487, "learning_rate": 4.9393750000000005e-06, "loss": 0.0292, "step": 9200 }, { "epoch": 4.221967963386728, "grad_norm": 3.233245849609375, "learning_rate": 4.9315625e-06, "loss": 0.0218, "step": 9225 }, { "epoch": 4.233409610983982, "grad_norm": 3.3073842525482178, "learning_rate": 4.92375e-06, "loss": 0.0301, "step": 9250 }, { "epoch": 4.244851258581235, "grad_norm": 3.375771999359131, "learning_rate": 4.9159375e-06, "loss": 0.0243, "step": 9275 }, { "epoch": 4.25629290617849, "grad_norm": 2.0122270584106445, "learning_rate": 4.9081250000000005e-06, "loss": 0.0274, "step": 9300 }, { "epoch": 4.267734553775743, "grad_norm": 4.077823638916016, "learning_rate": 4.9003125e-06, "loss": 0.0186, "step": 9325 }, { "epoch": 4.279176201372998, "grad_norm": 2.2521536350250244, "learning_rate": 4.8925e-06, "loss": 0.0305, "step": 9350 }, { "epoch": 4.290617848970252, "grad_norm": 2.4649903774261475, "learning_rate": 4.8846875e-06, "loss": 0.0218, "step": 9375 }, { "epoch": 4.302059496567506, "grad_norm": 2.196124792098999, "learning_rate": 4.876875e-06, "loss": 0.0342, "step": 9400 }, { "epoch": 4.31350114416476, "grad_norm": 2.979475975036621, "learning_rate": 4.8690625000000004e-06, "loss": 0.0242, "step": 9425 }, { "epoch": 4.324942791762013, "grad_norm": 7.853615760803223, "learning_rate": 4.86125e-06, "loss": 0.03, "step": 9450 }, { "epoch": 4.336384439359268, "grad_norm": 1.174206256866455, "learning_rate": 4.8534375e-06, "loss": 0.0195, "step": 9475 }, { "epoch": 4.3478260869565215, "grad_norm": 3.248457193374634, "learning_rate": 4.845625e-06, "loss": 0.0377, "step": 9500 }, { "epoch": 4.359267734553776, "grad_norm": 1.723983645439148, "learning_rate": 4.8378125000000005e-06, "loss": 0.0266, "step": 9525 }, { "epoch": 4.37070938215103, "grad_norm": 8.764972686767578, "learning_rate": 4.83e-06, "loss": 0.0431, "step": 9550 }, { "epoch": 4.382151029748284, "grad_norm": 1.3390229940414429, "learning_rate": 4.8221875e-06, "loss": 0.0176, "step": 9575 }, { "epoch": 4.393592677345538, "grad_norm": 4.078359603881836, "learning_rate": 4.814375e-06, "loss": 0.0338, "step": 9600 }, { "epoch": 4.4050343249427915, "grad_norm": 2.31992244720459, "learning_rate": 4.806562500000001e-06, "loss": 0.0202, "step": 9625 }, { "epoch": 4.416475972540046, "grad_norm": 4.521693229675293, "learning_rate": 4.7987500000000004e-06, "loss": 0.0302, "step": 9650 }, { "epoch": 4.4279176201373, "grad_norm": 5.699862480163574, "learning_rate": 4.7909375e-06, "loss": 0.0262, "step": 9675 }, { "epoch": 4.439359267734554, "grad_norm": 2.659482955932617, "learning_rate": 4.783125e-06, "loss": 0.0252, "step": 9700 }, { "epoch": 4.450800915331808, "grad_norm": 9.114242553710938, "learning_rate": 4.7753125e-06, "loss": 0.0292, "step": 9725 }, { "epoch": 4.462242562929061, "grad_norm": 2.6791884899139404, "learning_rate": 4.7675000000000005e-06, "loss": 0.0304, "step": 9750 }, { "epoch": 4.473684210526316, "grad_norm": 3.7395987510681152, "learning_rate": 4.7596875e-06, "loss": 0.0266, "step": 9775 }, { "epoch": 4.48512585812357, "grad_norm": 1.6912297010421753, "learning_rate": 4.751875e-06, "loss": 0.0446, "step": 9800 }, { "epoch": 4.496567505720824, "grad_norm": 3.4774134159088135, "learning_rate": 4.7440625e-06, "loss": 0.0266, "step": 9825 }, { "epoch": 4.508009153318078, "grad_norm": 3.602811098098755, "learning_rate": 4.736250000000001e-06, "loss": 0.0343, "step": 9850 }, { "epoch": 4.519450800915331, "grad_norm": 3.7448768615722656, "learning_rate": 4.7284374999999996e-06, "loss": 0.0241, "step": 9875 }, { "epoch": 4.530892448512586, "grad_norm": 2.9017863273620605, "learning_rate": 4.720625e-06, "loss": 0.0313, "step": 9900 }, { "epoch": 4.5423340961098395, "grad_norm": 2.8648183345794678, "learning_rate": 4.7128125e-06, "loss": 0.0224, "step": 9925 }, { "epoch": 4.553775743707094, "grad_norm": 2.749983549118042, "learning_rate": 4.705000000000001e-06, "loss": 0.0326, "step": 9950 }, { "epoch": 4.565217391304348, "grad_norm": 0.6364901661872864, "learning_rate": 4.6971875000000005e-06, "loss": 0.0192, "step": 9975 }, { "epoch": 4.576659038901602, "grad_norm": 0.9560081362724304, "learning_rate": 4.689375e-06, "loss": 0.0272, "step": 10000 }, { "epoch": 4.576659038901602, "eval_loss": 0.1832965463399887, "eval_runtime": 5483.9142, "eval_samples_per_second": 1.736, "eval_steps_per_second": 0.217, "eval_wer": 0.10381071222883471, "step": 10000 }, { "epoch": 4.588100686498856, "grad_norm": 4.101663589477539, "learning_rate": 4.6815625e-06, "loss": 0.0217, "step": 10025 }, { "epoch": 4.5995423340961095, "grad_norm": 2.0789246559143066, "learning_rate": 4.67375e-06, "loss": 0.0289, "step": 10050 }, { "epoch": 4.610983981693364, "grad_norm": 4.012986660003662, "learning_rate": 4.665937500000001e-06, "loss": 0.0272, "step": 10075 }, { "epoch": 4.622425629290618, "grad_norm": 2.1427500247955322, "learning_rate": 4.658125e-06, "loss": 0.0299, "step": 10100 }, { "epoch": 4.633867276887872, "grad_norm": 6.746981620788574, "learning_rate": 4.6503125e-06, "loss": 0.0222, "step": 10125 }, { "epoch": 4.645308924485126, "grad_norm": 0.5219627022743225, "learning_rate": 4.6425e-06, "loss": 0.0311, "step": 10150 }, { "epoch": 4.65675057208238, "grad_norm": 2.274179697036743, "learning_rate": 4.634687500000001e-06, "loss": 0.0189, "step": 10175 }, { "epoch": 4.668192219679634, "grad_norm": 1.2450759410858154, "learning_rate": 4.6271875e-06, "loss": 0.0319, "step": 10200 }, { "epoch": 4.679633867276888, "grad_norm": 7.748188018798828, "learning_rate": 4.619375e-06, "loss": 0.0249, "step": 10225 }, { "epoch": 4.691075514874142, "grad_norm": 2.7961182594299316, "learning_rate": 4.6115625e-06, "loss": 0.0367, "step": 10250 }, { "epoch": 4.702517162471396, "grad_norm": 1.6634931564331055, "learning_rate": 4.603750000000001e-06, "loss": 0.0193, "step": 10275 }, { "epoch": 4.71395881006865, "grad_norm": 3.7419142723083496, "learning_rate": 4.5959374999999996e-06, "loss": 0.0346, "step": 10300 }, { "epoch": 4.725400457665904, "grad_norm": 11.037392616271973, "learning_rate": 4.588125e-06, "loss": 0.0251, "step": 10325 }, { "epoch": 4.7368421052631575, "grad_norm": 2.88437819480896, "learning_rate": 4.5803125e-06, "loss": 0.0386, "step": 10350 }, { "epoch": 4.748283752860412, "grad_norm": 3.546595335006714, "learning_rate": 4.572500000000001e-06, "loss": 0.0184, "step": 10375 }, { "epoch": 4.759725400457666, "grad_norm": 1.2331124544143677, "learning_rate": 4.5646875000000005e-06, "loss": 0.0322, "step": 10400 }, { "epoch": 4.77116704805492, "grad_norm": 1.0554383993148804, "learning_rate": 4.556875e-06, "loss": 0.02, "step": 10425 }, { "epoch": 4.782608695652174, "grad_norm": 0.897574245929718, "learning_rate": 4.5490625e-06, "loss": 0.028, "step": 10450 }, { "epoch": 4.7940503432494275, "grad_norm": 3.45660138130188, "learning_rate": 4.541250000000001e-06, "loss": 0.0196, "step": 10475 }, { "epoch": 4.805491990846682, "grad_norm": 4.324442386627197, "learning_rate": 4.533437500000001e-06, "loss": 0.0329, "step": 10500 }, { "epoch": 4.816933638443936, "grad_norm": 7.328847885131836, "learning_rate": 4.525625e-06, "loss": 0.0286, "step": 10525 }, { "epoch": 4.82837528604119, "grad_norm": 0.6618658900260925, "learning_rate": 4.5178125e-06, "loss": 0.0257, "step": 10550 }, { "epoch": 4.839816933638444, "grad_norm": 3.4053359031677246, "learning_rate": 4.51e-06, "loss": 0.0188, "step": 10575 }, { "epoch": 4.851258581235698, "grad_norm": 1.5626916885375977, "learning_rate": 4.502187500000001e-06, "loss": 0.0388, "step": 10600 }, { "epoch": 4.862700228832952, "grad_norm": 2.1022493839263916, "learning_rate": 4.494375e-06, "loss": 0.0228, "step": 10625 }, { "epoch": 4.874141876430206, "grad_norm": 1.5179095268249512, "learning_rate": 4.4865625e-06, "loss": 0.0446, "step": 10650 }, { "epoch": 4.88558352402746, "grad_norm": 3.457355499267578, "learning_rate": 4.47875e-06, "loss": 0.0213, "step": 10675 }, { "epoch": 4.897025171624714, "grad_norm": 1.2291810512542725, "learning_rate": 4.470937500000001e-06, "loss": 0.0308, "step": 10700 }, { "epoch": 4.908466819221968, "grad_norm": 4.664734363555908, "learning_rate": 4.463125e-06, "loss": 0.0187, "step": 10725 }, { "epoch": 4.919908466819222, "grad_norm": 2.5962226390838623, "learning_rate": 4.4553125e-06, "loss": 0.0348, "step": 10750 }, { "epoch": 4.931350114416476, "grad_norm": 1.759839415550232, "learning_rate": 4.4475e-06, "loss": 0.0223, "step": 10775 }, { "epoch": 4.94279176201373, "grad_norm": 3.38629412651062, "learning_rate": 4.439687500000001e-06, "loss": 0.0439, "step": 10800 }, { "epoch": 4.954233409610984, "grad_norm": 0.6326417326927185, "learning_rate": 4.431875e-06, "loss": 0.0204, "step": 10825 }, { "epoch": 4.965675057208238, "grad_norm": 3.3679585456848145, "learning_rate": 4.4240625000000005e-06, "loss": 0.0322, "step": 10850 }, { "epoch": 4.977116704805492, "grad_norm": 3.8049206733703613, "learning_rate": 4.41625e-06, "loss": 0.0243, "step": 10875 }, { "epoch": 4.988558352402746, "grad_norm": 2.581594944000244, "learning_rate": 4.4084375e-06, "loss": 0.0332, "step": 10900 }, { "epoch": 5.0, "grad_norm": 3.7988698482513428, "learning_rate": 4.400625e-06, "loss": 0.0275, "step": 10925 }, { "epoch": 5.011441647597254, "grad_norm": 2.600945234298706, "learning_rate": 4.3928125e-06, "loss": 0.0131, "step": 10950 }, { "epoch": 5.022883295194508, "grad_norm": 8.007580757141113, "learning_rate": 4.385e-06, "loss": 0.0133, "step": 10975 }, { "epoch": 5.034324942791762, "grad_norm": 3.332880735397339, "learning_rate": 4.3771875e-06, "loss": 0.0149, "step": 11000 }, { "epoch": 5.045766590389016, "grad_norm": 15.387001037597656, "learning_rate": 4.369375000000001e-06, "loss": 0.0207, "step": 11025 }, { "epoch": 5.05720823798627, "grad_norm": 4.535585403442383, "learning_rate": 4.3615625e-06, "loss": 0.0199, "step": 11050 }, { "epoch": 5.068649885583524, "grad_norm": 4.385061740875244, "learning_rate": 4.3537500000000005e-06, "loss": 0.0248, "step": 11075 }, { "epoch": 5.080091533180778, "grad_norm": 2.6416256427764893, "learning_rate": 4.3459375e-06, "loss": 0.0129, "step": 11100 }, { "epoch": 5.091533180778032, "grad_norm": 2.3068976402282715, "learning_rate": 4.338125000000001e-06, "loss": 0.0271, "step": 11125 }, { "epoch": 5.102974828375286, "grad_norm": 5.006783485412598, "learning_rate": 4.3303125e-06, "loss": 0.0156, "step": 11150 }, { "epoch": 5.11441647597254, "grad_norm": 8.350238800048828, "learning_rate": 4.3225000000000005e-06, "loss": 0.0152, "step": 11175 }, { "epoch": 5.125858123569794, "grad_norm": 3.0252091884613037, "learning_rate": 4.3146875e-06, "loss": 0.0142, "step": 11200 }, { "epoch": 5.137299771167048, "grad_norm": 17.71987533569336, "learning_rate": 4.306875e-06, "loss": 0.0261, "step": 11225 }, { "epoch": 5.148741418764302, "grad_norm": 0.3432025611400604, "learning_rate": 4.2990625e-06, "loss": 0.0141, "step": 11250 }, { "epoch": 5.160183066361556, "grad_norm": 3.693594455718994, "learning_rate": 4.29125e-06, "loss": 0.0284, "step": 11275 }, { "epoch": 5.17162471395881, "grad_norm": 2.1907877922058105, "learning_rate": 4.2834375000000004e-06, "loss": 0.0103, "step": 11300 }, { "epoch": 5.183066361556064, "grad_norm": 5.526587963104248, "learning_rate": 4.275625e-06, "loss": 0.0254, "step": 11325 }, { "epoch": 5.194508009153318, "grad_norm": 5.500503063201904, "learning_rate": 4.2678125e-06, "loss": 0.0108, "step": 11350 }, { "epoch": 5.2059496567505725, "grad_norm": 7.589873790740967, "learning_rate": 4.26e-06, "loss": 0.0237, "step": 11375 }, { "epoch": 5.217391304347826, "grad_norm": 2.3860023021698, "learning_rate": 4.2521875000000005e-06, "loss": 0.0186, "step": 11400 }, { "epoch": 5.22883295194508, "grad_norm": 4.4563307762146, "learning_rate": 4.244375e-06, "loss": 0.0217, "step": 11425 }, { "epoch": 5.240274599542334, "grad_norm": 3.0135042667388916, "learning_rate": 4.2365625e-06, "loss": 0.0155, "step": 11450 }, { "epoch": 5.251716247139588, "grad_norm": 1.9074753522872925, "learning_rate": 4.22875e-06, "loss": 0.0251, "step": 11475 }, { "epoch": 5.2631578947368425, "grad_norm": 3.4067037105560303, "learning_rate": 4.220937500000001e-06, "loss": 0.0119, "step": 11500 }, { "epoch": 5.274599542334096, "grad_norm": 2.8989033699035645, "learning_rate": 4.2131250000000004e-06, "loss": 0.0188, "step": 11525 }, { "epoch": 5.28604118993135, "grad_norm": 2.8292016983032227, "learning_rate": 4.2053125e-06, "loss": 0.0159, "step": 11550 }, { "epoch": 5.297482837528604, "grad_norm": 6.212322235107422, "learning_rate": 4.1975e-06, "loss": 0.0233, "step": 11575 }, { "epoch": 5.308924485125858, "grad_norm": 0.9031001925468445, "learning_rate": 4.1896875e-06, "loss": 0.0146, "step": 11600 }, { "epoch": 5.320366132723112, "grad_norm": 7.238254070281982, "learning_rate": 4.1818750000000005e-06, "loss": 0.0158, "step": 11625 }, { "epoch": 5.331807780320366, "grad_norm": 2.4615652561187744, "learning_rate": 4.1740625e-06, "loss": 0.0122, "step": 11650 }, { "epoch": 5.34324942791762, "grad_norm": 3.7764835357666016, "learning_rate": 4.16625e-06, "loss": 0.0221, "step": 11675 }, { "epoch": 5.354691075514874, "grad_norm": 1.053728699684143, "learning_rate": 4.1584375e-06, "loss": 0.0186, "step": 11700 }, { "epoch": 5.366132723112128, "grad_norm": 7.169757843017578, "learning_rate": 4.150625000000001e-06, "loss": 0.0198, "step": 11725 }, { "epoch": 5.377574370709382, "grad_norm": 0.351454496383667, "learning_rate": 4.1428125e-06, "loss": 0.0121, "step": 11750 }, { "epoch": 5.389016018306636, "grad_norm": 4.189688682556152, "learning_rate": 4.135e-06, "loss": 0.0245, "step": 11775 }, { "epoch": 5.4004576659038905, "grad_norm": 3.2234408855438232, "learning_rate": 4.1271875e-06, "loss": 0.0149, "step": 11800 }, { "epoch": 5.411899313501144, "grad_norm": 7.368442058563232, "learning_rate": 4.119375000000001e-06, "loss": 0.0185, "step": 11825 }, { "epoch": 5.423340961098398, "grad_norm": 4.849565505981445, "learning_rate": 4.1115625000000005e-06, "loss": 0.0155, "step": 11850 }, { "epoch": 5.434782608695652, "grad_norm": 8.018900871276855, "learning_rate": 4.10375e-06, "loss": 0.0272, "step": 11875 }, { "epoch": 5.446224256292906, "grad_norm": 3.36999773979187, "learning_rate": 4.0959375e-06, "loss": 0.0168, "step": 11900 }, { "epoch": 5.4576659038901605, "grad_norm": 8.504731178283691, "learning_rate": 4.088125e-06, "loss": 0.0187, "step": 11925 }, { "epoch": 5.469107551487414, "grad_norm": 1.719467282295227, "learning_rate": 4.080312500000001e-06, "loss": 0.0213, "step": 11950 }, { "epoch": 5.480549199084669, "grad_norm": 7.953880310058594, "learning_rate": 4.0724999999999995e-06, "loss": 0.0209, "step": 11975 }, { "epoch": 5.491990846681922, "grad_norm": 3.389706611633301, "learning_rate": 4.0646875e-06, "loss": 0.0151, "step": 12000 }, { "epoch": 5.503432494279176, "grad_norm": 6.694132328033447, "learning_rate": 4.056875e-06, "loss": 0.0202, "step": 12025 }, { "epoch": 5.51487414187643, "grad_norm": 2.6016061305999756, "learning_rate": 4.049062500000001e-06, "loss": 0.0172, "step": 12050 }, { "epoch": 5.526315789473684, "grad_norm": 5.176774024963379, "learning_rate": 4.04125e-06, "loss": 0.0281, "step": 12075 }, { "epoch": 5.537757437070939, "grad_norm": 1.5615991353988647, "learning_rate": 4.0334375e-06, "loss": 0.0128, "step": 12100 }, { "epoch": 5.549199084668192, "grad_norm": 4.4115376472473145, "learning_rate": 4.025625e-06, "loss": 0.0278, "step": 12125 }, { "epoch": 5.560640732265446, "grad_norm": 0.18079940974712372, "learning_rate": 4.017812500000001e-06, "loss": 0.0129, "step": 12150 }, { "epoch": 5.5720823798627, "grad_norm": 1.855873942375183, "learning_rate": 4.01e-06, "loss": 0.024, "step": 12175 }, { "epoch": 5.583524027459954, "grad_norm": 3.256206512451172, "learning_rate": 4.0021875e-06, "loss": 0.0152, "step": 12200 }, { "epoch": 5.5949656750572085, "grad_norm": 6.555622100830078, "learning_rate": 3.994375e-06, "loss": 0.0261, "step": 12225 }, { "epoch": 5.606407322654462, "grad_norm": 3.024768352508545, "learning_rate": 3.9865625e-06, "loss": 0.0107, "step": 12250 }, { "epoch": 5.617848970251716, "grad_norm": 2.868229389190674, "learning_rate": 3.978750000000001e-06, "loss": 0.0255, "step": 12275 }, { "epoch": 5.62929061784897, "grad_norm": 5.408130168914795, "learning_rate": 3.9709375e-06, "loss": 0.0121, "step": 12300 }, { "epoch": 5.640732265446224, "grad_norm": 8.534693717956543, "learning_rate": 3.963125e-06, "loss": 0.0245, "step": 12325 }, { "epoch": 5.6521739130434785, "grad_norm": 0.8499470353126526, "learning_rate": 3.9553125e-06, "loss": 0.0114, "step": 12350 }, { "epoch": 5.663615560640732, "grad_norm": 1.4523133039474487, "learning_rate": 3.947500000000001e-06, "loss": 0.0251, "step": 12375 }, { "epoch": 5.675057208237987, "grad_norm": 1.7775627374649048, "learning_rate": 3.9396875e-06, "loss": 0.0191, "step": 12400 }, { "epoch": 5.68649885583524, "grad_norm": 4.454090118408203, "learning_rate": 3.931875e-06, "loss": 0.0195, "step": 12425 }, { "epoch": 5.697940503432494, "grad_norm": 1.4271085262298584, "learning_rate": 3.9240625e-06, "loss": 0.0138, "step": 12450 }, { "epoch": 5.709382151029748, "grad_norm": 9.393786430358887, "learning_rate": 3.916250000000001e-06, "loss": 0.0273, "step": 12475 }, { "epoch": 5.720823798627002, "grad_norm": 3.8522861003875732, "learning_rate": 3.9084375e-06, "loss": 0.0197, "step": 12500 }, { "epoch": 5.732265446224257, "grad_norm": 5.545007705688477, "learning_rate": 3.9006250000000005e-06, "loss": 0.0206, "step": 12525 }, { "epoch": 5.74370709382151, "grad_norm": 2.3138723373413086, "learning_rate": 3.8928125e-06, "loss": 0.0135, "step": 12550 }, { "epoch": 5.755148741418765, "grad_norm": 6.381573677062988, "learning_rate": 3.885e-06, "loss": 0.0264, "step": 12575 }, { "epoch": 5.766590389016018, "grad_norm": 2.9235124588012695, "learning_rate": 3.8771875e-06, "loss": 0.014, "step": 12600 }, { "epoch": 5.778032036613272, "grad_norm": 9.768881797790527, "learning_rate": 3.869375e-06, "loss": 0.0237, "step": 12625 }, { "epoch": 5.7894736842105265, "grad_norm": 2.5591719150543213, "learning_rate": 3.8615625e-06, "loss": 0.0129, "step": 12650 }, { "epoch": 5.80091533180778, "grad_norm": 8.308956146240234, "learning_rate": 3.85375e-06, "loss": 0.0174, "step": 12675 }, { "epoch": 5.812356979405035, "grad_norm": 1.5682135820388794, "learning_rate": 3.8459375e-06, "loss": 0.0134, "step": 12700 }, { "epoch": 5.823798627002288, "grad_norm": 5.886026859283447, "learning_rate": 3.838125e-06, "loss": 0.0221, "step": 12725 }, { "epoch": 5.835240274599542, "grad_norm": 5.910558223724365, "learning_rate": 3.8303125000000004e-06, "loss": 0.0163, "step": 12750 }, { "epoch": 5.8466819221967965, "grad_norm": 6.313669204711914, "learning_rate": 3.8225e-06, "loss": 0.0222, "step": 12775 }, { "epoch": 5.85812356979405, "grad_norm": 4.852139949798584, "learning_rate": 3.8146875e-06, "loss": 0.013, "step": 12800 }, { "epoch": 5.869565217391305, "grad_norm": 6.2581963539123535, "learning_rate": 3.806875e-06, "loss": 0.0169, "step": 12825 }, { "epoch": 5.881006864988558, "grad_norm": 5.26676607131958, "learning_rate": 3.7990625e-06, "loss": 0.0159, "step": 12850 }, { "epoch": 5.892448512585812, "grad_norm": 0.4348423182964325, "learning_rate": 3.7912500000000003e-06, "loss": 0.0276, "step": 12875 }, { "epoch": 5.9038901601830664, "grad_norm": 0.7607429623603821, "learning_rate": 3.7834375000000006e-06, "loss": 0.017, "step": 12900 }, { "epoch": 5.91533180778032, "grad_norm": 4.235646724700928, "learning_rate": 3.775625e-06, "loss": 0.0277, "step": 12925 }, { "epoch": 5.926773455377575, "grad_norm": 0.33068224787712097, "learning_rate": 3.7678125e-06, "loss": 0.0177, "step": 12950 }, { "epoch": 5.938215102974828, "grad_norm": 8.94812297821045, "learning_rate": 3.7600000000000004e-06, "loss": 0.0189, "step": 12975 }, { "epoch": 5.949656750572083, "grad_norm": 3.01238751411438, "learning_rate": 3.7521875000000007e-06, "loss": 0.0135, "step": 13000 }, { "epoch": 5.961098398169336, "grad_norm": 6.622538089752197, "learning_rate": 3.744375e-06, "loss": 0.021, "step": 13025 }, { "epoch": 5.97254004576659, "grad_norm": 5.920276641845703, "learning_rate": 3.7365625000000003e-06, "loss": 0.0208, "step": 13050 }, { "epoch": 5.983981693363845, "grad_norm": 1.632116436958313, "learning_rate": 3.7287500000000005e-06, "loss": 0.0206, "step": 13075 }, { "epoch": 5.995423340961098, "grad_norm": 6.903181076049805, "learning_rate": 3.7209375000000003e-06, "loss": 0.0155, "step": 13100 }, { "epoch": 6.006864988558353, "grad_norm": 0.6914149522781372, "learning_rate": 3.7134375e-06, "loss": 0.0194, "step": 13125 }, { "epoch": 6.018306636155606, "grad_norm": 6.280193328857422, "learning_rate": 3.705625e-06, "loss": 0.0168, "step": 13150 }, { "epoch": 6.02974828375286, "grad_norm": 0.3250044286251068, "learning_rate": 3.6978125000000004e-06, "loss": 0.0273, "step": 13175 }, { "epoch": 6.0411899313501145, "grad_norm": 2.4546151161193848, "learning_rate": 3.6900000000000002e-06, "loss": 0.0144, "step": 13200 }, { "epoch": 6.052631578947368, "grad_norm": 1.1390433311462402, "learning_rate": 3.6821875e-06, "loss": 0.0184, "step": 13225 }, { "epoch": 6.064073226544623, "grad_norm": 1.7204899787902832, "learning_rate": 3.674375e-06, "loss": 0.0092, "step": 13250 }, { "epoch": 6.075514874141876, "grad_norm": 1.4025479555130005, "learning_rate": 3.6665625e-06, "loss": 0.024, "step": 13275 }, { "epoch": 6.086956521739131, "grad_norm": 0.4468977153301239, "learning_rate": 3.6587500000000003e-06, "loss": 0.0139, "step": 13300 }, { "epoch": 6.0983981693363845, "grad_norm": 0.18809981644153595, "learning_rate": 3.6509374999999997e-06, "loss": 0.011, "step": 13325 }, { "epoch": 6.109839816933638, "grad_norm": 3.639801502227783, "learning_rate": 3.643125e-06, "loss": 0.0116, "step": 13350 }, { "epoch": 6.121281464530893, "grad_norm": 0.617262601852417, "learning_rate": 3.6353125e-06, "loss": 0.0127, "step": 13375 }, { "epoch": 6.132723112128146, "grad_norm": 3.9678919315338135, "learning_rate": 3.6275000000000004e-06, "loss": 0.0131, "step": 13400 }, { "epoch": 6.144164759725401, "grad_norm": 1.4072645902633667, "learning_rate": 3.6196875000000007e-06, "loss": 0.0144, "step": 13425 }, { "epoch": 6.155606407322654, "grad_norm": 6.396816253662109, "learning_rate": 3.611875e-06, "loss": 0.0094, "step": 13450 }, { "epoch": 6.167048054919908, "grad_norm": 2.001157760620117, "learning_rate": 3.6040625000000003e-06, "loss": 0.0189, "step": 13475 }, { "epoch": 6.178489702517163, "grad_norm": 4.379182815551758, "learning_rate": 3.5962500000000005e-06, "loss": 0.0135, "step": 13500 }, { "epoch": 6.189931350114416, "grad_norm": 1.7968692779541016, "learning_rate": 3.5884375000000003e-06, "loss": 0.0197, "step": 13525 }, { "epoch": 6.201372997711671, "grad_norm": 0.30796533823013306, "learning_rate": 3.580625e-06, "loss": 0.0091, "step": 13550 }, { "epoch": 6.212814645308924, "grad_norm": 0.5206483006477356, "learning_rate": 3.5728125e-06, "loss": 0.0193, "step": 13575 }, { "epoch": 6.224256292906179, "grad_norm": 0.16793321073055267, "learning_rate": 3.565e-06, "loss": 0.011, "step": 13600 }, { "epoch": 6.2356979405034325, "grad_norm": 3.4714410305023193, "learning_rate": 3.5571875000000004e-06, "loss": 0.0163, "step": 13625 }, { "epoch": 6.247139588100686, "grad_norm": 2.5397183895111084, "learning_rate": 3.549375e-06, "loss": 0.0098, "step": 13650 }, { "epoch": 6.258581235697941, "grad_norm": 0.8021149039268494, "learning_rate": 3.5415625e-06, "loss": 0.0261, "step": 13675 }, { "epoch": 6.270022883295194, "grad_norm": 2.6485466957092285, "learning_rate": 3.5337500000000003e-06, "loss": 0.0116, "step": 13700 }, { "epoch": 6.281464530892449, "grad_norm": 0.5701060891151428, "learning_rate": 3.5259375000000005e-06, "loss": 0.0144, "step": 13725 }, { "epoch": 6.2929061784897025, "grad_norm": 0.46138399839401245, "learning_rate": 3.518125e-06, "loss": 0.0074, "step": 13750 }, { "epoch": 6.304347826086957, "grad_norm": 2.842442512512207, "learning_rate": 3.5103125e-06, "loss": 0.0194, "step": 13775 }, { "epoch": 6.315789473684211, "grad_norm": 0.676575243473053, "learning_rate": 3.5025000000000003e-06, "loss": 0.0085, "step": 13800 }, { "epoch": 6.327231121281464, "grad_norm": 1.1983352899551392, "learning_rate": 3.4946875000000006e-06, "loss": 0.0118, "step": 13825 }, { "epoch": 6.338672768878719, "grad_norm": 3.670888900756836, "learning_rate": 3.486875e-06, "loss": 0.0126, "step": 13850 }, { "epoch": 6.350114416475972, "grad_norm": 0.25366395711898804, "learning_rate": 3.4790625e-06, "loss": 0.0154, "step": 13875 }, { "epoch": 6.361556064073227, "grad_norm": 3.8486428260803223, "learning_rate": 3.47125e-06, "loss": 0.0117, "step": 13900 }, { "epoch": 6.372997711670481, "grad_norm": 1.135246753692627, "learning_rate": 3.4634375000000002e-06, "loss": 0.0156, "step": 13925 }, { "epoch": 6.384439359267734, "grad_norm": 1.3957738876342773, "learning_rate": 3.4556249999999996e-06, "loss": 0.0191, "step": 13950 }, { "epoch": 6.395881006864989, "grad_norm": 0.1521887332201004, "learning_rate": 3.4478125e-06, "loss": 0.0161, "step": 13975 }, { "epoch": 6.407322654462242, "grad_norm": 0.3171142041683197, "learning_rate": 3.44e-06, "loss": 0.0106, "step": 14000 }, { "epoch": 6.418764302059497, "grad_norm": 1.6162185668945312, "learning_rate": 3.4321875000000003e-06, "loss": 0.016, "step": 14025 }, { "epoch": 6.4302059496567505, "grad_norm": 3.2044668197631836, "learning_rate": 3.4243750000000006e-06, "loss": 0.0088, "step": 14050 }, { "epoch": 6.441647597254004, "grad_norm": 16.820634841918945, "learning_rate": 3.4165625e-06, "loss": 0.0094, "step": 14075 }, { "epoch": 6.453089244851259, "grad_norm": 2.0577335357666016, "learning_rate": 3.40875e-06, "loss": 0.0113, "step": 14100 }, { "epoch": 6.464530892448512, "grad_norm": 2.2943193912506104, "learning_rate": 3.4009375000000004e-06, "loss": 0.0136, "step": 14125 }, { "epoch": 6.475972540045767, "grad_norm": 3.4963417053222656, "learning_rate": 3.3931250000000007e-06, "loss": 0.0138, "step": 14150 }, { "epoch": 6.4874141876430205, "grad_norm": 0.4365079402923584, "learning_rate": 3.3853125e-06, "loss": 0.0131, "step": 14175 }, { "epoch": 6.498855835240275, "grad_norm": 3.457284450531006, "learning_rate": 3.3775000000000003e-06, "loss": 0.0115, "step": 14200 }, { "epoch": 6.510297482837529, "grad_norm": 0.40817612409591675, "learning_rate": 3.3696875e-06, "loss": 0.0181, "step": 14225 }, { "epoch": 6.521739130434782, "grad_norm": 1.6476938724517822, "learning_rate": 3.3618750000000003e-06, "loss": 0.0097, "step": 14250 }, { "epoch": 6.533180778032037, "grad_norm": 0.25183960795402527, "learning_rate": 3.3540624999999997e-06, "loss": 0.0217, "step": 14275 }, { "epoch": 6.54462242562929, "grad_norm": 1.9451416730880737, "learning_rate": 3.34625e-06, "loss": 0.0097, "step": 14300 }, { "epoch": 6.556064073226545, "grad_norm": 4.131077289581299, "learning_rate": 3.3384375e-06, "loss": 0.0183, "step": 14325 }, { "epoch": 6.567505720823799, "grad_norm": 0.5438878536224365, "learning_rate": 3.3306250000000004e-06, "loss": 0.0145, "step": 14350 }, { "epoch": 6.578947368421053, "grad_norm": 0.5045881867408752, "learning_rate": 3.3228125e-06, "loss": 0.0172, "step": 14375 }, { "epoch": 6.590389016018307, "grad_norm": 0.4585050344467163, "learning_rate": 3.315e-06, "loss": 0.0126, "step": 14400 }, { "epoch": 6.60183066361556, "grad_norm": 0.8127508759498596, "learning_rate": 3.3071875000000003e-06, "loss": 0.0151, "step": 14425 }, { "epoch": 6.613272311212815, "grad_norm": 1.7050418853759766, "learning_rate": 3.2993750000000005e-06, "loss": 0.009, "step": 14450 }, { "epoch": 6.6247139588100685, "grad_norm": 2.609612464904785, "learning_rate": 3.2915625e-06, "loss": 0.014, "step": 14475 }, { "epoch": 6.636155606407323, "grad_norm": 0.10997484624385834, "learning_rate": 3.28375e-06, "loss": 0.0094, "step": 14500 }, { "epoch": 6.647597254004577, "grad_norm": 5.7333173751831055, "learning_rate": 3.2759375000000003e-06, "loss": 0.0154, "step": 14525 }, { "epoch": 6.65903890160183, "grad_norm": 2.564332962036133, "learning_rate": 3.268125e-06, "loss": 0.0155, "step": 14550 }, { "epoch": 6.670480549199085, "grad_norm": 4.814024448394775, "learning_rate": 3.2603125e-06, "loss": 0.0179, "step": 14575 }, { "epoch": 6.6819221967963385, "grad_norm": 1.1423102617263794, "learning_rate": 3.2525e-06, "loss": 0.0096, "step": 14600 }, { "epoch": 6.693363844393593, "grad_norm": 3.1103646755218506, "learning_rate": 3.2446875e-06, "loss": 0.0164, "step": 14625 }, { "epoch": 6.704805491990847, "grad_norm": 1.021026372909546, "learning_rate": 3.2368750000000002e-06, "loss": 0.0075, "step": 14650 }, { "epoch": 6.7162471395881, "grad_norm": 0.4085843861103058, "learning_rate": 3.2290625000000005e-06, "loss": 0.0167, "step": 14675 }, { "epoch": 6.727688787185355, "grad_norm": 0.8315253257751465, "learning_rate": 3.22125e-06, "loss": 0.0085, "step": 14700 }, { "epoch": 6.739130434782608, "grad_norm": 2.2641472816467285, "learning_rate": 3.2134375e-06, "loss": 0.0209, "step": 14725 }, { "epoch": 6.750572082379863, "grad_norm": 6.209875583648682, "learning_rate": 3.2056250000000003e-06, "loss": 0.0088, "step": 14750 }, { "epoch": 6.762013729977117, "grad_norm": 0.5979451537132263, "learning_rate": 3.1978125000000006e-06, "loss": 0.0179, "step": 14775 }, { "epoch": 6.77345537757437, "grad_norm": 2.0241873264312744, "learning_rate": 3.19e-06, "loss": 0.0103, "step": 14800 }, { "epoch": 6.784897025171625, "grad_norm": 0.17280341684818268, "learning_rate": 3.1821875e-06, "loss": 0.0227, "step": 14825 }, { "epoch": 6.796338672768878, "grad_norm": 0.12923689186573029, "learning_rate": 3.1743750000000004e-06, "loss": 0.0089, "step": 14850 }, { "epoch": 6.807780320366133, "grad_norm": 0.43938687443733215, "learning_rate": 3.1665625000000002e-06, "loss": 0.0209, "step": 14875 }, { "epoch": 6.8192219679633865, "grad_norm": 0.2618752419948578, "learning_rate": 3.15875e-06, "loss": 0.0109, "step": 14900 }, { "epoch": 6.830663615560641, "grad_norm": 3.6829495429992676, "learning_rate": 3.1509375000000003e-06, "loss": 0.0188, "step": 14925 }, { "epoch": 6.842105263157895, "grad_norm": 4.100037574768066, "learning_rate": 3.143125e-06, "loss": 0.0103, "step": 14950 }, { "epoch": 6.853546910755149, "grad_norm": 3.131757974624634, "learning_rate": 3.1353125000000003e-06, "loss": 0.0199, "step": 14975 }, { "epoch": 6.864988558352403, "grad_norm": 2.9150915145874023, "learning_rate": 3.1274999999999997e-06, "loss": 0.0125, "step": 15000 }, { "epoch": 6.864988558352403, "eval_loss": 0.20364722609519958, "eval_runtime": 5377.8653, "eval_samples_per_second": 1.771, "eval_steps_per_second": 0.221, "eval_wer": 0.09913931016829845, "step": 15000 }, { "epoch": 6.8764302059496565, "grad_norm": 3.365856885910034, "learning_rate": 3.1196875e-06, "loss": 0.0218, "step": 15025 }, { "epoch": 6.887871853546911, "grad_norm": 0.5725418329238892, "learning_rate": 3.111875e-06, "loss": 0.0073, "step": 15050 }, { "epoch": 6.899313501144165, "grad_norm": 1.2597202062606812, "learning_rate": 3.1040625e-06, "loss": 0.0155, "step": 15075 }, { "epoch": 6.910755148741419, "grad_norm": 4.598627090454102, "learning_rate": 3.0962500000000002e-06, "loss": 0.0098, "step": 15100 }, { "epoch": 6.922196796338673, "grad_norm": 0.7468834519386292, "learning_rate": 3.0884375e-06, "loss": 0.0193, "step": 15125 }, { "epoch": 6.933638443935926, "grad_norm": 4.2396464347839355, "learning_rate": 3.0806250000000003e-06, "loss": 0.0102, "step": 15150 }, { "epoch": 6.945080091533181, "grad_norm": 2.1572372913360596, "learning_rate": 3.0728125e-06, "loss": 0.02, "step": 15175 }, { "epoch": 6.956521739130435, "grad_norm": 2.8561244010925293, "learning_rate": 3.0650000000000003e-06, "loss": 0.0096, "step": 15200 }, { "epoch": 6.967963386727689, "grad_norm": 0.29279011487960815, "learning_rate": 3.0571875e-06, "loss": 0.0175, "step": 15225 }, { "epoch": 6.979405034324943, "grad_norm": 1.120410442352295, "learning_rate": 3.0493750000000003e-06, "loss": 0.0117, "step": 15250 }, { "epoch": 6.990846681922196, "grad_norm": 2.2122271060943604, "learning_rate": 3.0415625e-06, "loss": 0.0133, "step": 15275 }, { "epoch": 7.002288329519451, "grad_norm": 1.584817886352539, "learning_rate": 3.03375e-06, "loss": 0.0088, "step": 15300 }, { "epoch": 7.0137299771167045, "grad_norm": 0.8186086416244507, "learning_rate": 3.0259375e-06, "loss": 0.0064, "step": 15325 }, { "epoch": 7.025171624713959, "grad_norm": 1.3982340097427368, "learning_rate": 3.018125e-06, "loss": 0.0173, "step": 15350 }, { "epoch": 7.036613272311213, "grad_norm": 0.623369574546814, "learning_rate": 3.0103125000000002e-06, "loss": 0.0072, "step": 15375 }, { "epoch": 7.048054919908467, "grad_norm": 0.9551053643226624, "learning_rate": 3.0025e-06, "loss": 0.0104, "step": 15400 }, { "epoch": 7.059496567505721, "grad_norm": 0.4638427495956421, "learning_rate": 2.9946875000000003e-06, "loss": 0.0074, "step": 15425 }, { "epoch": 7.0709382151029745, "grad_norm": 0.7113415598869324, "learning_rate": 2.986875e-06, "loss": 0.0121, "step": 15450 }, { "epoch": 7.082379862700229, "grad_norm": 1.5616718530654907, "learning_rate": 2.9790625000000003e-06, "loss": 0.0084, "step": 15475 }, { "epoch": 7.093821510297483, "grad_norm": 5.3068528175354, "learning_rate": 2.97125e-06, "loss": 0.0115, "step": 15500 }, { "epoch": 7.105263157894737, "grad_norm": 6.045175075531006, "learning_rate": 2.9634375000000004e-06, "loss": 0.007, "step": 15525 }, { "epoch": 7.116704805491991, "grad_norm": 0.12478996068239212, "learning_rate": 2.955625e-06, "loss": 0.0122, "step": 15550 }, { "epoch": 7.128146453089244, "grad_norm": 0.3576812744140625, "learning_rate": 2.9478125000000004e-06, "loss": 0.007, "step": 15575 }, { "epoch": 7.139588100686499, "grad_norm": 1.844315767288208, "learning_rate": 2.9400000000000002e-06, "loss": 0.0115, "step": 15600 }, { "epoch": 7.151029748283753, "grad_norm": 9.679522514343262, "learning_rate": 2.9321875e-06, "loss": 0.0087, "step": 15625 }, { "epoch": 7.162471395881007, "grad_norm": 0.7042823433876038, "learning_rate": 2.924375e-06, "loss": 0.0125, "step": 15650 }, { "epoch": 7.173913043478261, "grad_norm": 0.31098079681396484, "learning_rate": 2.9165625e-06, "loss": 0.0099, "step": 15675 }, { "epoch": 7.185354691075515, "grad_norm": 2.236577033996582, "learning_rate": 2.90875e-06, "loss": 0.0152, "step": 15700 }, { "epoch": 7.196796338672769, "grad_norm": 0.2922148108482361, "learning_rate": 2.9009375e-06, "loss": 0.0066, "step": 15725 }, { "epoch": 7.2082379862700225, "grad_norm": 3.0574800968170166, "learning_rate": 2.893125e-06, "loss": 0.0142, "step": 15750 }, { "epoch": 7.219679633867277, "grad_norm": 0.9282683730125427, "learning_rate": 2.8853125e-06, "loss": 0.007, "step": 15775 }, { "epoch": 7.231121281464531, "grad_norm": 1.8068912029266357, "learning_rate": 2.8775e-06, "loss": 0.0065, "step": 15800 }, { "epoch": 7.242562929061785, "grad_norm": 2.347912311553955, "learning_rate": 2.8696875000000002e-06, "loss": 0.0098, "step": 15825 }, { "epoch": 7.254004576659039, "grad_norm": 1.0899382829666138, "learning_rate": 2.861875e-06, "loss": 0.0143, "step": 15850 }, { "epoch": 7.2654462242562925, "grad_norm": 0.7211443185806274, "learning_rate": 2.8540625000000003e-06, "loss": 0.0057, "step": 15875 }, { "epoch": 7.276887871853547, "grad_norm": 1.6489779949188232, "learning_rate": 2.8462500000000005e-06, "loss": 0.0175, "step": 15900 }, { "epoch": 7.288329519450801, "grad_norm": 1.7807652950286865, "learning_rate": 2.8384375000000003e-06, "loss": 0.0074, "step": 15925 }, { "epoch": 7.299771167048055, "grad_norm": 2.1116390228271484, "learning_rate": 2.830625e-06, "loss": 0.0111, "step": 15950 }, { "epoch": 7.311212814645309, "grad_norm": 0.1760861575603485, "learning_rate": 2.8228125e-06, "loss": 0.0098, "step": 15975 }, { "epoch": 7.322654462242563, "grad_norm": 1.8475443124771118, "learning_rate": 2.815e-06, "loss": 0.0093, "step": 16000 }, { "epoch": 7.334096109839817, "grad_norm": 0.14219874143600464, "learning_rate": 2.8071875e-06, "loss": 0.0057, "step": 16025 }, { "epoch": 7.345537757437071, "grad_norm": 0.7175412774085999, "learning_rate": 2.799375e-06, "loss": 0.0195, "step": 16050 }, { "epoch": 7.356979405034325, "grad_norm": 0.1249295026063919, "learning_rate": 2.7915625e-06, "loss": 0.0085, "step": 16075 }, { "epoch": 7.368421052631579, "grad_norm": 1.440131664276123, "learning_rate": 2.7837500000000002e-06, "loss": 0.0176, "step": 16100 }, { "epoch": 7.379862700228833, "grad_norm": 0.3971177339553833, "learning_rate": 2.7759375e-06, "loss": 0.0073, "step": 16125 }, { "epoch": 7.391304347826087, "grad_norm": 3.1487958431243896, "learning_rate": 2.7681250000000003e-06, "loss": 0.0099, "step": 16150 }, { "epoch": 7.4027459954233406, "grad_norm": 4.346505165100098, "learning_rate": 2.7603125e-06, "loss": 0.0061, "step": 16175 }, { "epoch": 7.414187643020595, "grad_norm": 1.5192569494247437, "learning_rate": 2.7525000000000003e-06, "loss": 0.0195, "step": 16200 }, { "epoch": 7.425629290617849, "grad_norm": 2.2581429481506348, "learning_rate": 2.7446875e-06, "loss": 0.0071, "step": 16225 }, { "epoch": 7.437070938215103, "grad_norm": 1.6350927352905273, "learning_rate": 2.7368750000000004e-06, "loss": 0.0115, "step": 16250 }, { "epoch": 7.448512585812357, "grad_norm": 0.6200711727142334, "learning_rate": 2.7290625e-06, "loss": 0.0083, "step": 16275 }, { "epoch": 7.459954233409611, "grad_norm": 0.9812812805175781, "learning_rate": 2.72125e-06, "loss": 0.013, "step": 16300 }, { "epoch": 7.471395881006865, "grad_norm": 0.33793002367019653, "learning_rate": 2.7134375e-06, "loss": 0.0074, "step": 16325 }, { "epoch": 7.482837528604119, "grad_norm": 1.2503453493118286, "learning_rate": 2.705625e-06, "loss": 0.011, "step": 16350 }, { "epoch": 7.494279176201373, "grad_norm": 0.3572548031806946, "learning_rate": 2.6978125e-06, "loss": 0.0058, "step": 16375 }, { "epoch": 7.505720823798627, "grad_norm": 1.6965501308441162, "learning_rate": 2.69e-06, "loss": 0.0101, "step": 16400 }, { "epoch": 7.517162471395881, "grad_norm": 1.6538255214691162, "learning_rate": 2.6821875e-06, "loss": 0.0079, "step": 16425 }, { "epoch": 7.528604118993135, "grad_norm": 1.0653047561645508, "learning_rate": 2.674375e-06, "loss": 0.0247, "step": 16450 }, { "epoch": 7.540045766590389, "grad_norm": 0.19629091024398804, "learning_rate": 2.6665625e-06, "loss": 0.0093, "step": 16475 }, { "epoch": 7.551487414187643, "grad_norm": 0.48006531596183777, "learning_rate": 2.65875e-06, "loss": 0.0133, "step": 16500 }, { "epoch": 7.562929061784897, "grad_norm": 0.11753380298614502, "learning_rate": 2.6509375000000004e-06, "loss": 0.0052, "step": 16525 }, { "epoch": 7.574370709382151, "grad_norm": 0.975831151008606, "learning_rate": 2.6434375e-06, "loss": 0.0115, "step": 16550 }, { "epoch": 7.585812356979405, "grad_norm": 3.755007028579712, "learning_rate": 2.6356250000000003e-06, "loss": 0.0086, "step": 16575 }, { "epoch": 7.597254004576659, "grad_norm": 0.4555506408214569, "learning_rate": 2.6278125e-06, "loss": 0.0178, "step": 16600 }, { "epoch": 7.608695652173913, "grad_norm": 0.7911310791969299, "learning_rate": 2.6200000000000003e-06, "loss": 0.0098, "step": 16625 }, { "epoch": 7.620137299771167, "grad_norm": 1.7671183347702026, "learning_rate": 2.6121875e-06, "loss": 0.0178, "step": 16650 }, { "epoch": 7.631578947368421, "grad_norm": 3.4077210426330566, "learning_rate": 2.6043750000000004e-06, "loss": 0.0107, "step": 16675 }, { "epoch": 7.643020594965675, "grad_norm": 0.5805812478065491, "learning_rate": 2.5965625e-06, "loss": 0.014, "step": 16700 }, { "epoch": 7.654462242562929, "grad_norm": 0.3688525855541229, "learning_rate": 2.5887500000000004e-06, "loss": 0.0069, "step": 16725 }, { "epoch": 7.665903890160183, "grad_norm": 0.8044073581695557, "learning_rate": 2.5809375000000002e-06, "loss": 0.0158, "step": 16750 }, { "epoch": 7.6773455377574376, "grad_norm": 0.10179438441991806, "learning_rate": 2.573125e-06, "loss": 0.0083, "step": 16775 }, { "epoch": 7.688787185354691, "grad_norm": 1.3863494396209717, "learning_rate": 2.5653125e-06, "loss": 0.015, "step": 16800 }, { "epoch": 7.700228832951945, "grad_norm": 0.7638989090919495, "learning_rate": 2.5575e-06, "loss": 0.0102, "step": 16825 }, { "epoch": 7.711670480549199, "grad_norm": 1.0550347566604614, "learning_rate": 2.5496875e-06, "loss": 0.0142, "step": 16850 }, { "epoch": 7.723112128146453, "grad_norm": 6.330323219299316, "learning_rate": 2.541875e-06, "loss": 0.0099, "step": 16875 }, { "epoch": 7.7345537757437075, "grad_norm": 3.124359369277954, "learning_rate": 2.5340625e-06, "loss": 0.008, "step": 16900 }, { "epoch": 7.745995423340961, "grad_norm": 0.30625733733177185, "learning_rate": 2.52625e-06, "loss": 0.006, "step": 16925 }, { "epoch": 7.757437070938215, "grad_norm": 1.6125200986862183, "learning_rate": 2.5184375e-06, "loss": 0.0142, "step": 16950 }, { "epoch": 7.768878718535469, "grad_norm": 0.9505332708358765, "learning_rate": 2.510625e-06, "loss": 0.0062, "step": 16975 }, { "epoch": 7.780320366132723, "grad_norm": 3.020979404449463, "learning_rate": 2.5028125e-06, "loss": 0.0169, "step": 17000 }, { "epoch": 7.7917620137299775, "grad_norm": 0.600764274597168, "learning_rate": 2.4950000000000003e-06, "loss": 0.0064, "step": 17025 }, { "epoch": 7.803203661327231, "grad_norm": 1.8066636323928833, "learning_rate": 2.4871875000000005e-06, "loss": 0.0189, "step": 17050 }, { "epoch": 7.814645308924485, "grad_norm": 9.108110427856445, "learning_rate": 2.4793750000000003e-06, "loss": 0.0071, "step": 17075 }, { "epoch": 7.826086956521739, "grad_norm": 2.945603132247925, "learning_rate": 2.4715625e-06, "loss": 0.0089, "step": 17100 }, { "epoch": 7.837528604118993, "grad_norm": 0.26592642068862915, "learning_rate": 2.46375e-06, "loss": 0.0071, "step": 17125 }, { "epoch": 7.848970251716247, "grad_norm": 0.4964580833911896, "learning_rate": 2.4559375e-06, "loss": 0.0089, "step": 17150 }, { "epoch": 7.860411899313501, "grad_norm": 3.0700931549072266, "learning_rate": 2.448125e-06, "loss": 0.0129, "step": 17175 }, { "epoch": 7.871853546910755, "grad_norm": 1.6877917051315308, "learning_rate": 2.4403125e-06, "loss": 0.0187, "step": 17200 }, { "epoch": 7.883295194508009, "grad_norm": 0.2074529379606247, "learning_rate": 2.4325e-06, "loss": 0.0101, "step": 17225 }, { "epoch": 7.894736842105263, "grad_norm": 0.1611488312482834, "learning_rate": 2.4246875000000002e-06, "loss": 0.0163, "step": 17250 }, { "epoch": 7.906178489702517, "grad_norm": 21.526241302490234, "learning_rate": 2.416875e-06, "loss": 0.0096, "step": 17275 }, { "epoch": 7.917620137299771, "grad_norm": 1.5529296398162842, "learning_rate": 2.4090625000000003e-06, "loss": 0.0142, "step": 17300 }, { "epoch": 7.9290617848970255, "grad_norm": 0.151444673538208, "learning_rate": 2.40125e-06, "loss": 0.0066, "step": 17325 }, { "epoch": 7.940503432494279, "grad_norm": 1.1893980503082275, "learning_rate": 2.3934375000000003e-06, "loss": 0.0076, "step": 17350 }, { "epoch": 7.951945080091534, "grad_norm": 3.2917895317077637, "learning_rate": 2.385625e-06, "loss": 0.0082, "step": 17375 }, { "epoch": 7.963386727688787, "grad_norm": 1.4442905187606812, "learning_rate": 2.3778125000000004e-06, "loss": 0.0079, "step": 17400 }, { "epoch": 7.974828375286041, "grad_norm": 3.447230577468872, "learning_rate": 2.37e-06, "loss": 0.0117, "step": 17425 }, { "epoch": 7.9862700228832955, "grad_norm": 0.0969911515712738, "learning_rate": 2.3621875e-06, "loss": 0.0186, "step": 17450 }, { "epoch": 7.997711670480549, "grad_norm": 3.7749619483947754, "learning_rate": 2.354375e-06, "loss": 0.0064, "step": 17475 }, { "epoch": 8.009153318077804, "grad_norm": 0.26138579845428467, "learning_rate": 2.3465625e-06, "loss": 0.0155, "step": 17500 }, { "epoch": 8.020594965675057, "grad_norm": 1.6767950057983398, "learning_rate": 2.33875e-06, "loss": 0.0067, "step": 17525 }, { "epoch": 8.03203661327231, "grad_norm": 2.5147764682769775, "learning_rate": 2.3309375e-06, "loss": 0.0128, "step": 17550 }, { "epoch": 8.043478260869565, "grad_norm": 0.12537100911140442, "learning_rate": 2.323125e-06, "loss": 0.0071, "step": 17575 }, { "epoch": 8.05491990846682, "grad_norm": 7.097085952758789, "learning_rate": 2.3153125e-06, "loss": 0.0069, "step": 17600 }, { "epoch": 8.066361556064074, "grad_norm": 0.1975085735321045, "learning_rate": 2.3075e-06, "loss": 0.0063, "step": 17625 }, { "epoch": 8.077803203661327, "grad_norm": 2.5416934490203857, "learning_rate": 2.2996875e-06, "loss": 0.0067, "step": 17650 }, { "epoch": 8.08924485125858, "grad_norm": 0.9218634963035583, "learning_rate": 2.2918750000000004e-06, "loss": 0.0081, "step": 17675 }, { "epoch": 8.100686498855834, "grad_norm": 1.0671167373657227, "learning_rate": 2.2840625e-06, "loss": 0.0153, "step": 17700 }, { "epoch": 8.11212814645309, "grad_norm": 0.07980553060770035, "learning_rate": 2.2762500000000004e-06, "loss": 0.0121, "step": 17725 }, { "epoch": 8.123569794050344, "grad_norm": 0.6690077781677246, "learning_rate": 2.2684375000000003e-06, "loss": 0.0061, "step": 17750 }, { "epoch": 8.135011441647597, "grad_norm": 14.465129852294922, "learning_rate": 2.260625e-06, "loss": 0.0083, "step": 17775 }, { "epoch": 8.14645308924485, "grad_norm": 1.2041326761245728, "learning_rate": 2.2528125e-06, "loss": 0.0093, "step": 17800 }, { "epoch": 8.157894736842104, "grad_norm": 6.081242561340332, "learning_rate": 2.245e-06, "loss": 0.0068, "step": 17825 }, { "epoch": 8.16933638443936, "grad_norm": 1.041059970855713, "learning_rate": 2.2371875e-06, "loss": 0.012, "step": 17850 }, { "epoch": 8.180778032036613, "grad_norm": 2.5304393768310547, "learning_rate": 2.229375e-06, "loss": 0.0145, "step": 17875 }, { "epoch": 8.192219679633867, "grad_norm": 0.06209346652030945, "learning_rate": 2.2215625e-06, "loss": 0.0061, "step": 17900 }, { "epoch": 8.20366132723112, "grad_norm": 0.9206859469413757, "learning_rate": 2.21375e-06, "loss": 0.0116, "step": 17925 }, { "epoch": 8.215102974828376, "grad_norm": 1.2147653102874756, "learning_rate": 2.2059375e-06, "loss": 0.0067, "step": 17950 }, { "epoch": 8.22654462242563, "grad_norm": 0.1464168131351471, "learning_rate": 2.1981250000000002e-06, "loss": 0.0078, "step": 17975 }, { "epoch": 8.237986270022883, "grad_norm": 1.9483212232589722, "learning_rate": 2.1903125e-06, "loss": 0.0112, "step": 18000 }, { "epoch": 8.249427917620137, "grad_norm": 3.755913734436035, "learning_rate": 2.1825000000000003e-06, "loss": 0.0093, "step": 18025 }, { "epoch": 8.26086956521739, "grad_norm": 1.1962236166000366, "learning_rate": 2.1746875e-06, "loss": 0.007, "step": 18050 }, { "epoch": 8.272311212814646, "grad_norm": 1.5700790882110596, "learning_rate": 2.1668750000000003e-06, "loss": 0.0115, "step": 18075 }, { "epoch": 8.2837528604119, "grad_norm": 0.4117244780063629, "learning_rate": 2.1590625e-06, "loss": 0.0089, "step": 18100 }, { "epoch": 8.295194508009153, "grad_norm": 2.4972431659698486, "learning_rate": 2.15125e-06, "loss": 0.0051, "step": 18125 }, { "epoch": 8.306636155606407, "grad_norm": 0.6796992421150208, "learning_rate": 2.1434374999999998e-06, "loss": 0.01, "step": 18150 }, { "epoch": 8.31807780320366, "grad_norm": 1.471119999885559, "learning_rate": 2.135625e-06, "loss": 0.0051, "step": 18175 }, { "epoch": 8.329519450800916, "grad_norm": 0.32256969809532166, "learning_rate": 2.1278125e-06, "loss": 0.01, "step": 18200 }, { "epoch": 8.34096109839817, "grad_norm": 1.6328589916229248, "learning_rate": 2.12e-06, "loss": 0.0045, "step": 18225 }, { "epoch": 8.352402745995423, "grad_norm": 0.09561590105295181, "learning_rate": 2.1121875e-06, "loss": 0.0106, "step": 18250 }, { "epoch": 8.363844393592677, "grad_norm": 4.785649299621582, "learning_rate": 2.104375e-06, "loss": 0.0094, "step": 18275 }, { "epoch": 8.37528604118993, "grad_norm": 1.282939076423645, "learning_rate": 2.0965625000000003e-06, "loss": 0.0095, "step": 18300 }, { "epoch": 8.386727688787186, "grad_norm": 0.11949565261602402, "learning_rate": 2.08875e-06, "loss": 0.0082, "step": 18325 }, { "epoch": 8.39816933638444, "grad_norm": 0.13847044110298157, "learning_rate": 2.0809375000000004e-06, "loss": 0.01, "step": 18350 }, { "epoch": 8.409610983981693, "grad_norm": 0.5209012031555176, "learning_rate": 2.073125e-06, "loss": 0.0073, "step": 18375 }, { "epoch": 8.421052631578947, "grad_norm": 0.20265115797519684, "learning_rate": 2.0653125000000004e-06, "loss": 0.0115, "step": 18400 }, { "epoch": 8.4324942791762, "grad_norm": 0.9013919830322266, "learning_rate": 2.0575e-06, "loss": 0.0099, "step": 18425 }, { "epoch": 8.443935926773456, "grad_norm": 1.9062713384628296, "learning_rate": 2.0496875e-06, "loss": 0.0112, "step": 18450 }, { "epoch": 8.45537757437071, "grad_norm": 3.9823992252349854, "learning_rate": 2.041875e-06, "loss": 0.0066, "step": 18475 }, { "epoch": 8.466819221967963, "grad_norm": 0.3470716178417206, "learning_rate": 2.0340625e-06, "loss": 0.0049, "step": 18500 }, { "epoch": 8.478260869565217, "grad_norm": 1.912090539932251, "learning_rate": 2.02625e-06, "loss": 0.0072, "step": 18525 }, { "epoch": 8.48970251716247, "grad_norm": 2.102421522140503, "learning_rate": 2.0184375e-06, "loss": 0.0081, "step": 18550 }, { "epoch": 8.501144164759726, "grad_norm": 0.11117032915353775, "learning_rate": 2.010625e-06, "loss": 0.0057, "step": 18575 }, { "epoch": 8.51258581235698, "grad_norm": 0.21074584126472473, "learning_rate": 2.0028125e-06, "loss": 0.0103, "step": 18600 }, { "epoch": 8.524027459954233, "grad_norm": 0.12100081890821457, "learning_rate": 1.995e-06, "loss": 0.0038, "step": 18625 }, { "epoch": 8.535469107551487, "grad_norm": 0.08020362257957458, "learning_rate": 1.9871875e-06, "loss": 0.0077, "step": 18650 }, { "epoch": 8.546910755148742, "grad_norm": 2.3164095878601074, "learning_rate": 1.979375e-06, "loss": 0.0087, "step": 18675 }, { "epoch": 8.558352402745996, "grad_norm": 0.48693156242370605, "learning_rate": 1.9715625000000002e-06, "loss": 0.0079, "step": 18700 }, { "epoch": 8.56979405034325, "grad_norm": 0.06513327360153198, "learning_rate": 1.96375e-06, "loss": 0.0093, "step": 18725 }, { "epoch": 8.581235697940503, "grad_norm": 0.049644988030195236, "learning_rate": 1.9559375000000003e-06, "loss": 0.0122, "step": 18750 }, { "epoch": 8.592677345537757, "grad_norm": 0.46888506412506104, "learning_rate": 1.948125e-06, "loss": 0.0066, "step": 18775 }, { "epoch": 8.604118993135012, "grad_norm": 1.531025767326355, "learning_rate": 1.9403125000000003e-06, "loss": 0.0089, "step": 18800 }, { "epoch": 8.615560640732266, "grad_norm": 3.112746000289917, "learning_rate": 1.9325e-06, "loss": 0.0101, "step": 18825 }, { "epoch": 8.62700228832952, "grad_norm": 0.7876713871955872, "learning_rate": 1.9246875e-06, "loss": 0.0138, "step": 18850 }, { "epoch": 8.638443935926773, "grad_norm": 16.939899444580078, "learning_rate": 1.9168749999999998e-06, "loss": 0.0098, "step": 18875 }, { "epoch": 8.649885583524027, "grad_norm": 0.5488317608833313, "learning_rate": 1.9090625e-06, "loss": 0.0098, "step": 18900 }, { "epoch": 8.661327231121282, "grad_norm": 9.495604515075684, "learning_rate": 1.9012500000000002e-06, "loss": 0.0125, "step": 18925 }, { "epoch": 8.672768878718536, "grad_norm": 0.20728307962417603, "learning_rate": 1.8934375e-06, "loss": 0.0084, "step": 18950 }, { "epoch": 8.68421052631579, "grad_norm": 13.799328804016113, "learning_rate": 1.8856250000000003e-06, "loss": 0.0169, "step": 18975 }, { "epoch": 8.695652173913043, "grad_norm": 0.3443595767021179, "learning_rate": 1.8778125e-06, "loss": 0.0129, "step": 19000 }, { "epoch": 8.707093821510298, "grad_norm": 0.05261196941137314, "learning_rate": 1.8700000000000003e-06, "loss": 0.0081, "step": 19025 }, { "epoch": 8.718535469107552, "grad_norm": 1.2424705028533936, "learning_rate": 1.8621875000000001e-06, "loss": 0.0138, "step": 19050 }, { "epoch": 8.729977116704806, "grad_norm": 2.1272494792938232, "learning_rate": 1.8543750000000001e-06, "loss": 0.0087, "step": 19075 }, { "epoch": 8.74141876430206, "grad_norm": 0.19991889595985413, "learning_rate": 1.8465625e-06, "loss": 0.011, "step": 19100 }, { "epoch": 8.752860411899313, "grad_norm": 2.920642375946045, "learning_rate": 1.8387500000000002e-06, "loss": 0.0072, "step": 19125 }, { "epoch": 8.764302059496568, "grad_norm": 2.8176496028900146, "learning_rate": 1.8309375e-06, "loss": 0.0093, "step": 19150 }, { "epoch": 8.775743707093822, "grad_norm": 2.0975587368011475, "learning_rate": 1.8231250000000002e-06, "loss": 0.0071, "step": 19175 }, { "epoch": 8.787185354691076, "grad_norm": 0.34839051961898804, "learning_rate": 1.8153125e-06, "loss": 0.0061, "step": 19200 }, { "epoch": 8.79862700228833, "grad_norm": 1.3293085098266602, "learning_rate": 1.8075000000000003e-06, "loss": 0.0088, "step": 19225 }, { "epoch": 8.810068649885583, "grad_norm": 0.2044825553894043, "learning_rate": 1.7996875e-06, "loss": 0.0085, "step": 19250 }, { "epoch": 8.821510297482838, "grad_norm": 0.24040932953357697, "learning_rate": 1.791875e-06, "loss": 0.0073, "step": 19275 }, { "epoch": 8.832951945080092, "grad_norm": 4.296323299407959, "learning_rate": 1.784375e-06, "loss": 0.0103, "step": 19300 }, { "epoch": 8.844393592677346, "grad_norm": 0.16996651887893677, "learning_rate": 1.7765625000000002e-06, "loss": 0.0046, "step": 19325 }, { "epoch": 8.8558352402746, "grad_norm": 7.846490859985352, "learning_rate": 1.76875e-06, "loss": 0.0091, "step": 19350 }, { "epoch": 8.867276887871853, "grad_norm": 6.62661600112915, "learning_rate": 1.7609375e-06, "loss": 0.0038, "step": 19375 }, { "epoch": 8.878718535469108, "grad_norm": 0.48901718854904175, "learning_rate": 1.7531250000000003e-06, "loss": 0.0084, "step": 19400 }, { "epoch": 8.890160183066362, "grad_norm": 0.669843852519989, "learning_rate": 1.7453125e-06, "loss": 0.0073, "step": 19425 }, { "epoch": 8.901601830663616, "grad_norm": 5.433796405792236, "learning_rate": 1.7375000000000003e-06, "loss": 0.0065, "step": 19450 }, { "epoch": 8.91304347826087, "grad_norm": 0.07918695360422134, "learning_rate": 1.7296875000000001e-06, "loss": 0.0056, "step": 19475 }, { "epoch": 8.924485125858123, "grad_norm": 0.07582997530698776, "learning_rate": 1.7218750000000001e-06, "loss": 0.0115, "step": 19500 }, { "epoch": 8.935926773455378, "grad_norm": 0.5449735522270203, "learning_rate": 1.7140625e-06, "loss": 0.0106, "step": 19525 }, { "epoch": 8.947368421052632, "grad_norm": 0.7232295274734497, "learning_rate": 1.7062500000000002e-06, "loss": 0.0059, "step": 19550 }, { "epoch": 8.958810068649885, "grad_norm": 4.808300495147705, "learning_rate": 1.6984375e-06, "loss": 0.0103, "step": 19575 }, { "epoch": 8.97025171624714, "grad_norm": 5.207526683807373, "learning_rate": 1.6906250000000002e-06, "loss": 0.0104, "step": 19600 }, { "epoch": 8.981693363844393, "grad_norm": 0.05005680024623871, "learning_rate": 1.6828125e-06, "loss": 0.0056, "step": 19625 }, { "epoch": 8.993135011441648, "grad_norm": 0.113949254155159, "learning_rate": 1.6750000000000003e-06, "loss": 0.0061, "step": 19650 }, { "epoch": 9.004576659038902, "grad_norm": 0.4994412064552307, "learning_rate": 1.6671875e-06, "loss": 0.0114, "step": 19675 }, { "epoch": 9.016018306636155, "grad_norm": 1.9782668352127075, "learning_rate": 1.659375e-06, "loss": 0.005, "step": 19700 }, { "epoch": 9.027459954233409, "grad_norm": 0.5642163157463074, "learning_rate": 1.6515625e-06, "loss": 0.0128, "step": 19725 }, { "epoch": 9.038901601830664, "grad_norm": 0.13898131251335144, "learning_rate": 1.6437500000000001e-06, "loss": 0.0057, "step": 19750 }, { "epoch": 9.050343249427918, "grad_norm": 0.932721734046936, "learning_rate": 1.6359375e-06, "loss": 0.0059, "step": 19775 }, { "epoch": 9.061784897025172, "grad_norm": 3.428349494934082, "learning_rate": 1.6281250000000002e-06, "loss": 0.0084, "step": 19800 }, { "epoch": 9.073226544622425, "grad_norm": 3.3652799129486084, "learning_rate": 1.6203125e-06, "loss": 0.0117, "step": 19825 }, { "epoch": 9.084668192219679, "grad_norm": 0.9847972393035889, "learning_rate": 1.6125e-06, "loss": 0.0039, "step": 19850 }, { "epoch": 9.096109839816934, "grad_norm": 0.21777117252349854, "learning_rate": 1.6046875e-06, "loss": 0.0077, "step": 19875 }, { "epoch": 9.107551487414188, "grad_norm": 0.054129473865032196, "learning_rate": 1.596875e-06, "loss": 0.0053, "step": 19900 }, { "epoch": 9.118993135011442, "grad_norm": 0.08982221782207489, "learning_rate": 1.5890624999999999e-06, "loss": 0.0065, "step": 19925 }, { "epoch": 9.130434782608695, "grad_norm": 0.7745457291603088, "learning_rate": 1.5812500000000001e-06, "loss": 0.0055, "step": 19950 }, { "epoch": 9.141876430205949, "grad_norm": 1.2945647239685059, "learning_rate": 1.5734375e-06, "loss": 0.0082, "step": 19975 }, { "epoch": 9.153318077803204, "grad_norm": 1.5046677589416504, "learning_rate": 1.5656250000000002e-06, "loss": 0.0057, "step": 20000 }, { "epoch": 9.153318077803204, "eval_loss": 0.2091646045446396, "eval_runtime": 5410.4377, "eval_samples_per_second": 1.76, "eval_steps_per_second": 0.22, "eval_wer": 0.0982754207461445, "step": 20000 }, { "epoch": 9.164759725400458, "grad_norm": 0.4961409866809845, "learning_rate": 1.5578125000000002e-06, "loss": 0.0101, "step": 20025 }, { "epoch": 9.176201372997712, "grad_norm": 0.1165071576833725, "learning_rate": 1.55e-06, "loss": 0.0049, "step": 20050 }, { "epoch": 9.187643020594965, "grad_norm": 4.693184852600098, "learning_rate": 1.5421875e-06, "loss": 0.0071, "step": 20075 }, { "epoch": 9.199084668192219, "grad_norm": 2.6000266075134277, "learning_rate": 1.534375e-06, "loss": 0.0042, "step": 20100 }, { "epoch": 9.210526315789474, "grad_norm": 0.3854849338531494, "learning_rate": 1.5265625e-06, "loss": 0.0047, "step": 20125 }, { "epoch": 9.221967963386728, "grad_norm": 0.0833137035369873, "learning_rate": 1.51875e-06, "loss": 0.0045, "step": 20150 }, { "epoch": 9.233409610983982, "grad_norm": 0.8595607876777649, "learning_rate": 1.5109375e-06, "loss": 0.0059, "step": 20175 }, { "epoch": 9.244851258581235, "grad_norm": 0.15006107091903687, "learning_rate": 1.5031250000000001e-06, "loss": 0.0035, "step": 20200 }, { "epoch": 9.256292906178489, "grad_norm": 0.9896960258483887, "learning_rate": 1.4953125e-06, "loss": 0.0064, "step": 20225 }, { "epoch": 9.267734553775744, "grad_norm": 6.129720211029053, "learning_rate": 1.4875e-06, "loss": 0.0043, "step": 20250 }, { "epoch": 9.279176201372998, "grad_norm": 0.558292031288147, "learning_rate": 1.4796875e-06, "loss": 0.0134, "step": 20275 }, { "epoch": 9.290617848970252, "grad_norm": 0.06486662477254868, "learning_rate": 1.471875e-06, "loss": 0.0048, "step": 20300 }, { "epoch": 9.302059496567505, "grad_norm": 0.475968599319458, "learning_rate": 1.4640625000000002e-06, "loss": 0.0102, "step": 20325 }, { "epoch": 9.31350114416476, "grad_norm": 1.7022502422332764, "learning_rate": 1.4562500000000002e-06, "loss": 0.0038, "step": 20350 }, { "epoch": 9.324942791762014, "grad_norm": 3.257913589477539, "learning_rate": 1.4484375e-06, "loss": 0.0123, "step": 20375 }, { "epoch": 9.336384439359268, "grad_norm": 0.739531397819519, "learning_rate": 1.440625e-06, "loss": 0.0066, "step": 20400 }, { "epoch": 9.347826086956522, "grad_norm": 0.6568574905395508, "learning_rate": 1.4328125e-06, "loss": 0.0081, "step": 20425 }, { "epoch": 9.359267734553775, "grad_norm": 0.15944476425647736, "learning_rate": 1.4250000000000001e-06, "loss": 0.0041, "step": 20450 }, { "epoch": 9.37070938215103, "grad_norm": 2.0112485885620117, "learning_rate": 1.4171875000000001e-06, "loss": 0.0114, "step": 20475 }, { "epoch": 9.382151029748284, "grad_norm": 6.05720853805542, "learning_rate": 1.4093750000000002e-06, "loss": 0.0074, "step": 20500 }, { "epoch": 9.393592677345538, "grad_norm": 5.496974468231201, "learning_rate": 1.4015625000000002e-06, "loss": 0.0071, "step": 20525 }, { "epoch": 9.405034324942791, "grad_norm": 0.04030141234397888, "learning_rate": 1.39375e-06, "loss": 0.003, "step": 20550 }, { "epoch": 9.416475972540045, "grad_norm": 2.8988194465637207, "learning_rate": 1.3859375e-06, "loss": 0.0073, "step": 20575 }, { "epoch": 9.4279176201373, "grad_norm": 3.288280963897705, "learning_rate": 1.378125e-06, "loss": 0.0038, "step": 20600 }, { "epoch": 9.439359267734554, "grad_norm": 0.5999969244003296, "learning_rate": 1.3703125e-06, "loss": 0.0072, "step": 20625 }, { "epoch": 9.450800915331808, "grad_norm": 0.05087543651461601, "learning_rate": 1.3625e-06, "loss": 0.0047, "step": 20650 }, { "epoch": 9.462242562929061, "grad_norm": 0.31114375591278076, "learning_rate": 1.3546875e-06, "loss": 0.0118, "step": 20675 }, { "epoch": 9.473684210526315, "grad_norm": 0.05578223988413811, "learning_rate": 1.3468750000000001e-06, "loss": 0.0068, "step": 20700 }, { "epoch": 9.48512585812357, "grad_norm": 0.2985018789768219, "learning_rate": 1.3390625e-06, "loss": 0.0135, "step": 20725 }, { "epoch": 9.496567505720824, "grad_norm": 8.602644920349121, "learning_rate": 1.33125e-06, "loss": 0.0064, "step": 20750 }, { "epoch": 9.508009153318078, "grad_norm": 0.8129026293754578, "learning_rate": 1.3234375e-06, "loss": 0.0074, "step": 20775 }, { "epoch": 9.519450800915331, "grad_norm": 13.631547927856445, "learning_rate": 1.315625e-06, "loss": 0.0054, "step": 20800 }, { "epoch": 9.530892448512585, "grad_norm": 1.1775178909301758, "learning_rate": 1.3078125e-06, "loss": 0.0147, "step": 20825 }, { "epoch": 9.54233409610984, "grad_norm": 3.9796142578125, "learning_rate": 1.3e-06, "loss": 0.0072, "step": 20850 }, { "epoch": 9.553775743707094, "grad_norm": 0.5741226077079773, "learning_rate": 1.2921875e-06, "loss": 0.0093, "step": 20875 }, { "epoch": 9.565217391304348, "grad_norm": 0.49431973695755005, "learning_rate": 1.284375e-06, "loss": 0.0042, "step": 20900 }, { "epoch": 9.576659038901601, "grad_norm": 0.24845191836357117, "learning_rate": 1.2765625e-06, "loss": 0.0059, "step": 20925 }, { "epoch": 9.588100686498855, "grad_norm": 0.3911905884742737, "learning_rate": 1.2687500000000001e-06, "loss": 0.0066, "step": 20950 }, { "epoch": 9.59954233409611, "grad_norm": 4.166600704193115, "learning_rate": 1.2609375000000002e-06, "loss": 0.0102, "step": 20975 }, { "epoch": 9.610983981693364, "grad_norm": 1.9600577354431152, "learning_rate": 1.2531250000000002e-06, "loss": 0.0055, "step": 21000 }, { "epoch": 9.622425629290618, "grad_norm": 0.053466055542230606, "learning_rate": 1.2453125000000002e-06, "loss": 0.0081, "step": 21025 }, { "epoch": 9.633867276887871, "grad_norm": 0.10502848774194717, "learning_rate": 1.2375000000000002e-06, "loss": 0.0046, "step": 21050 }, { "epoch": 9.645308924485127, "grad_norm": 0.6094481348991394, "learning_rate": 1.2296875e-06, "loss": 0.0081, "step": 21075 }, { "epoch": 9.65675057208238, "grad_norm": 0.2322821468114853, "learning_rate": 1.221875e-06, "loss": 0.0064, "step": 21100 }, { "epoch": 9.668192219679634, "grad_norm": 0.5858094692230225, "learning_rate": 1.2140625e-06, "loss": 0.0099, "step": 21125 }, { "epoch": 9.679633867276888, "grad_norm": 0.4338737428188324, "learning_rate": 1.20625e-06, "loss": 0.0053, "step": 21150 }, { "epoch": 9.691075514874141, "grad_norm": 2.8088247776031494, "learning_rate": 1.1984375000000001e-06, "loss": 0.0096, "step": 21175 }, { "epoch": 9.702517162471397, "grad_norm": 0.27802830934524536, "learning_rate": 1.1906250000000001e-06, "loss": 0.005, "step": 21200 }, { "epoch": 9.71395881006865, "grad_norm": 0.3727511167526245, "learning_rate": 1.1828125000000002e-06, "loss": 0.012, "step": 21225 }, { "epoch": 9.725400457665904, "grad_norm": 0.17081360518932343, "learning_rate": 1.175e-06, "loss": 0.0059, "step": 21250 }, { "epoch": 9.736842105263158, "grad_norm": 2.140308380126953, "learning_rate": 1.1671875e-06, "loss": 0.0146, "step": 21275 }, { "epoch": 9.748283752860411, "grad_norm": 0.04215441644191742, "learning_rate": 1.159375e-06, "loss": 0.0056, "step": 21300 }, { "epoch": 9.759725400457667, "grad_norm": 1.3280128240585327, "learning_rate": 1.1515625e-06, "loss": 0.0049, "step": 21325 }, { "epoch": 9.77116704805492, "grad_norm": 0.8858660459518433, "learning_rate": 1.14375e-06, "loss": 0.0061, "step": 21350 }, { "epoch": 9.782608695652174, "grad_norm": 2.4523348808288574, "learning_rate": 1.1359375e-06, "loss": 0.0074, "step": 21375 }, { "epoch": 9.794050343249427, "grad_norm": 0.11217644065618515, "learning_rate": 1.128125e-06, "loss": 0.0037, "step": 21400 }, { "epoch": 9.805491990846681, "grad_norm": 0.23124535381793976, "learning_rate": 1.120625e-06, "loss": 0.008, "step": 21425 }, { "epoch": 9.816933638443937, "grad_norm": 4.284655570983887, "learning_rate": 1.1128125000000002e-06, "loss": 0.0083, "step": 21450 }, { "epoch": 9.82837528604119, "grad_norm": 1.4181989431381226, "learning_rate": 1.1050000000000002e-06, "loss": 0.0068, "step": 21475 }, { "epoch": 9.839816933638444, "grad_norm": 0.043931469321250916, "learning_rate": 1.0971875e-06, "loss": 0.0056, "step": 21500 }, { "epoch": 9.851258581235697, "grad_norm": 1.3745752573013306, "learning_rate": 1.089375e-06, "loss": 0.0053, "step": 21525 }, { "epoch": 9.862700228832953, "grad_norm": 4.101833820343018, "learning_rate": 1.0815625e-06, "loss": 0.005, "step": 21550 }, { "epoch": 9.874141876430206, "grad_norm": 0.13216155767440796, "learning_rate": 1.07375e-06, "loss": 0.0046, "step": 21575 }, { "epoch": 9.88558352402746, "grad_norm": 0.1495962142944336, "learning_rate": 1.0659375000000001e-06, "loss": 0.006, "step": 21600 }, { "epoch": 9.897025171624714, "grad_norm": 6.025510787963867, "learning_rate": 1.0581250000000001e-06, "loss": 0.0096, "step": 21625 }, { "epoch": 9.908466819221967, "grad_norm": 1.0735716819763184, "learning_rate": 1.0503125000000002e-06, "loss": 0.006, "step": 21650 }, { "epoch": 9.919908466819223, "grad_norm": 0.1290319263935089, "learning_rate": 1.0425e-06, "loss": 0.0062, "step": 21675 }, { "epoch": 9.931350114416476, "grad_norm": 0.3358752429485321, "learning_rate": 1.0346875e-06, "loss": 0.0072, "step": 21700 }, { "epoch": 9.94279176201373, "grad_norm": 0.3810190260410309, "learning_rate": 1.026875e-06, "loss": 0.0054, "step": 21725 }, { "epoch": 9.954233409610984, "grad_norm": 0.14147138595581055, "learning_rate": 1.0190625e-06, "loss": 0.0041, "step": 21750 }, { "epoch": 9.965675057208237, "grad_norm": 0.8347612619400024, "learning_rate": 1.01125e-06, "loss": 0.0087, "step": 21775 }, { "epoch": 9.977116704805493, "grad_norm": 0.0881999209523201, "learning_rate": 1.0034375e-06, "loss": 0.0072, "step": 21800 }, { "epoch": 9.988558352402746, "grad_norm": 0.05394062027335167, "learning_rate": 9.95625e-07, "loss": 0.0085, "step": 21825 }, { "epoch": 10.0, "grad_norm": 5.682041645050049, "learning_rate": 9.878125000000001e-07, "loss": 0.008, "step": 21850 }, { "epoch": 10.011441647597254, "grad_norm": 0.030163856223225594, "learning_rate": 9.8e-07, "loss": 0.0035, "step": 21875 }, { "epoch": 10.022883295194507, "grad_norm": 1.5702613592147827, "learning_rate": 9.721875e-07, "loss": 0.0059, "step": 21900 }, { "epoch": 10.034324942791763, "grad_norm": 1.831508994102478, "learning_rate": 9.64375e-07, "loss": 0.007, "step": 21925 }, { "epoch": 10.045766590389016, "grad_norm": 3.454017162322998, "learning_rate": 9.565625e-07, "loss": 0.007, "step": 21950 }, { "epoch": 10.05720823798627, "grad_norm": 0.42849200963974, "learning_rate": 9.4875e-07, "loss": 0.0029, "step": 21975 }, { "epoch": 10.068649885583524, "grad_norm": 10.296757698059082, "learning_rate": 9.409374999999999e-07, "loss": 0.012, "step": 22000 }, { "epoch": 10.080091533180777, "grad_norm": 2.301107168197632, "learning_rate": 9.33125e-07, "loss": 0.0064, "step": 22025 }, { "epoch": 10.091533180778033, "grad_norm": 0.2854618728160858, "learning_rate": 9.253125e-07, "loss": 0.009, "step": 22050 }, { "epoch": 10.102974828375286, "grad_norm": 3.725808620452881, "learning_rate": 9.175000000000001e-07, "loss": 0.0039, "step": 22075 }, { "epoch": 10.11441647597254, "grad_norm": 5.867912769317627, "learning_rate": 9.096875000000001e-07, "loss": 0.0078, "step": 22100 }, { "epoch": 10.125858123569794, "grad_norm": 3.184048652648926, "learning_rate": 9.018750000000002e-07, "loss": 0.0051, "step": 22125 }, { "epoch": 10.137299771167047, "grad_norm": 4.497511386871338, "learning_rate": 8.940625000000001e-07, "loss": 0.0089, "step": 22150 }, { "epoch": 10.148741418764303, "grad_norm": 0.0537056103348732, "learning_rate": 8.862500000000001e-07, "loss": 0.0027, "step": 22175 }, { "epoch": 10.160183066361556, "grad_norm": 3.5934197902679443, "learning_rate": 8.784375000000001e-07, "loss": 0.009, "step": 22200 }, { "epoch": 10.17162471395881, "grad_norm": 0.23545417189598083, "learning_rate": 8.706250000000001e-07, "loss": 0.0058, "step": 22225 }, { "epoch": 10.183066361556063, "grad_norm": 5.6573004722595215, "learning_rate": 8.628125e-07, "loss": 0.0082, "step": 22250 }, { "epoch": 10.194508009153319, "grad_norm": 1.168822169303894, "learning_rate": 8.550000000000001e-07, "loss": 0.0032, "step": 22275 }, { "epoch": 10.205949656750573, "grad_norm": 4.02138090133667, "learning_rate": 8.471875000000001e-07, "loss": 0.011, "step": 22300 }, { "epoch": 10.217391304347826, "grad_norm": 0.18906021118164062, "learning_rate": 8.39375e-07, "loss": 0.0042, "step": 22325 }, { "epoch": 10.22883295194508, "grad_norm": 4.549426078796387, "learning_rate": 8.315625e-07, "loss": 0.0127, "step": 22350 }, { "epoch": 10.240274599542333, "grad_norm": 0.021799422800540924, "learning_rate": 8.237500000000001e-07, "loss": 0.0056, "step": 22375 }, { "epoch": 10.251716247139589, "grad_norm": 6.450737953186035, "learning_rate": 8.159375000000001e-07, "loss": 0.0037, "step": 22400 }, { "epoch": 10.263157894736842, "grad_norm": 0.04190767928957939, "learning_rate": 8.08125e-07, "loss": 0.0087, "step": 22425 }, { "epoch": 10.274599542334096, "grad_norm": 5.531692028045654, "learning_rate": 8.003125e-07, "loss": 0.0076, "step": 22450 }, { "epoch": 10.28604118993135, "grad_norm": 0.1595929116010666, "learning_rate": 7.925e-07, "loss": 0.0043, "step": 22475 }, { "epoch": 10.297482837528603, "grad_norm": 2.6428611278533936, "learning_rate": 7.846875000000001e-07, "loss": 0.0107, "step": 22500 }, { "epoch": 10.308924485125859, "grad_norm": 0.043632976710796356, "learning_rate": 7.76875e-07, "loss": 0.0055, "step": 22525 }, { "epoch": 10.320366132723112, "grad_norm": 4.209513187408447, "learning_rate": 7.690625000000001e-07, "loss": 0.014, "step": 22550 }, { "epoch": 10.331807780320366, "grad_norm": 0.08739282190799713, "learning_rate": 7.612500000000001e-07, "loss": 0.0054, "step": 22575 }, { "epoch": 10.34324942791762, "grad_norm": 5.106118202209473, "learning_rate": 7.534375e-07, "loss": 0.0048, "step": 22600 }, { "epoch": 10.354691075514873, "grad_norm": 1.0995253324508667, "learning_rate": 7.456250000000001e-07, "loss": 0.0048, "step": 22625 }, { "epoch": 10.366132723112129, "grad_norm": 4.488454341888428, "learning_rate": 7.378125000000001e-07, "loss": 0.0083, "step": 22650 }, { "epoch": 10.377574370709382, "grad_norm": 0.04510258138179779, "learning_rate": 7.3e-07, "loss": 0.0054, "step": 22675 }, { "epoch": 10.389016018306636, "grad_norm": 12.380441665649414, "learning_rate": 7.221875e-07, "loss": 0.0053, "step": 22700 }, { "epoch": 10.40045766590389, "grad_norm": 0.06387301534414291, "learning_rate": 7.14375e-07, "loss": 0.0036, "step": 22725 }, { "epoch": 10.411899313501145, "grad_norm": 5.258612155914307, "learning_rate": 7.065625000000001e-07, "loss": 0.0044, "step": 22750 }, { "epoch": 10.423340961098399, "grad_norm": 1.0784777402877808, "learning_rate": 6.9875e-07, "loss": 0.0023, "step": 22775 }, { "epoch": 10.434782608695652, "grad_norm": 9.012079238891602, "learning_rate": 6.909375e-07, "loss": 0.0085, "step": 22800 }, { "epoch": 10.446224256292906, "grad_norm": 0.2832639515399933, "learning_rate": 6.83125e-07, "loss": 0.005, "step": 22825 }, { "epoch": 10.45766590389016, "grad_norm": 1.8746007680892944, "learning_rate": 6.753124999999999e-07, "loss": 0.0067, "step": 22850 }, { "epoch": 10.469107551487415, "grad_norm": 0.7013452649116516, "learning_rate": 6.675000000000001e-07, "loss": 0.0035, "step": 22875 }, { "epoch": 10.480549199084669, "grad_norm": 4.465462684631348, "learning_rate": 6.596875000000001e-07, "loss": 0.0048, "step": 22900 }, { "epoch": 10.491990846681922, "grad_norm": 3.2530734539031982, "learning_rate": 6.51875e-07, "loss": 0.0062, "step": 22925 }, { "epoch": 10.503432494279176, "grad_norm": 11.731400489807129, "learning_rate": 6.440625e-07, "loss": 0.0157, "step": 22950 }, { "epoch": 10.51487414187643, "grad_norm": 0.5493900775909424, "learning_rate": 6.3625e-07, "loss": 0.0039, "step": 22975 }, { "epoch": 10.526315789473685, "grad_norm": 4.6817240715026855, "learning_rate": 6.284375000000001e-07, "loss": 0.0092, "step": 23000 }, { "epoch": 10.537757437070939, "grad_norm": 1.2315994501113892, "learning_rate": 6.20625e-07, "loss": 0.0044, "step": 23025 }, { "epoch": 10.549199084668192, "grad_norm": 0.23386216163635254, "learning_rate": 6.128125e-07, "loss": 0.0079, "step": 23050 }, { "epoch": 10.560640732265446, "grad_norm": 0.387458860874176, "learning_rate": 6.05e-07, "loss": 0.0032, "step": 23075 }, { "epoch": 10.5720823798627, "grad_norm": 3.660433769226074, "learning_rate": 5.971875e-07, "loss": 0.0036, "step": 23100 }, { "epoch": 10.583524027459955, "grad_norm": 2.7630832195281982, "learning_rate": 5.89375e-07, "loss": 0.0068, "step": 23125 }, { "epoch": 10.594965675057209, "grad_norm": 12.062283515930176, "learning_rate": 5.815625e-07, "loss": 0.0101, "step": 23150 }, { "epoch": 10.606407322654462, "grad_norm": 0.07887246459722519, "learning_rate": 5.737500000000001e-07, "loss": 0.0036, "step": 23175 }, { "epoch": 10.617848970251716, "grad_norm": 0.8604605793952942, "learning_rate": 5.659375e-07, "loss": 0.008, "step": 23200 }, { "epoch": 10.62929061784897, "grad_norm": 0.21621568500995636, "learning_rate": 5.581250000000001e-07, "loss": 0.0036, "step": 23225 }, { "epoch": 10.640732265446225, "grad_norm": 4.55475378036499, "learning_rate": 5.503125000000001e-07, "loss": 0.0049, "step": 23250 }, { "epoch": 10.652173913043478, "grad_norm": 0.07743289321660995, "learning_rate": 5.425e-07, "loss": 0.0022, "step": 23275 }, { "epoch": 10.663615560640732, "grad_norm": 2.8529226779937744, "learning_rate": 5.346875e-07, "loss": 0.0035, "step": 23300 }, { "epoch": 10.675057208237986, "grad_norm": 1.4317353963851929, "learning_rate": 5.26875e-07, "loss": 0.0046, "step": 23325 }, { "epoch": 10.68649885583524, "grad_norm": 4.198233127593994, "learning_rate": 5.190625000000001e-07, "loss": 0.0039, "step": 23350 }, { "epoch": 10.697940503432495, "grad_norm": 1.182904601097107, "learning_rate": 5.1125e-07, "loss": 0.0053, "step": 23375 }, { "epoch": 10.709382151029748, "grad_norm": 7.1977105140686035, "learning_rate": 5.034375e-07, "loss": 0.0103, "step": 23400 }, { "epoch": 10.720823798627002, "grad_norm": 0.051739439368247986, "learning_rate": 4.95625e-07, "loss": 0.0044, "step": 23425 }, { "epoch": 10.732265446224256, "grad_norm": 6.68768310546875, "learning_rate": 4.878125e-07, "loss": 0.0084, "step": 23450 }, { "epoch": 10.743707093821511, "grad_norm": 0.04325169697403908, "learning_rate": 4.8e-07, "loss": 0.0054, "step": 23475 }, { "epoch": 10.755148741418765, "grad_norm": 7.35172700881958, "learning_rate": 4.721875000000001e-07, "loss": 0.0173, "step": 23500 }, { "epoch": 10.766590389016018, "grad_norm": 1.9433234930038452, "learning_rate": 4.6437500000000005e-07, "loss": 0.0032, "step": 23525 }, { "epoch": 10.778032036613272, "grad_norm": 9.553720474243164, "learning_rate": 4.565625e-07, "loss": 0.0047, "step": 23550 }, { "epoch": 10.789473684210526, "grad_norm": 0.2103455811738968, "learning_rate": 4.4875000000000004e-07, "loss": 0.0047, "step": 23575 }, { "epoch": 10.800915331807781, "grad_norm": 0.601758599281311, "learning_rate": 4.409375e-07, "loss": 0.0091, "step": 23600 }, { "epoch": 10.812356979405035, "grad_norm": 0.04723617061972618, "learning_rate": 4.3312500000000004e-07, "loss": 0.0043, "step": 23625 }, { "epoch": 10.823798627002288, "grad_norm": 0.1756594479084015, "learning_rate": 4.253125e-07, "loss": 0.0068, "step": 23650 }, { "epoch": 10.835240274599542, "grad_norm": 0.13647685945034027, "learning_rate": 4.175e-07, "loss": 0.0035, "step": 23675 }, { "epoch": 10.846681922196796, "grad_norm": 6.657943248748779, "learning_rate": 4.096875e-07, "loss": 0.0086, "step": 23700 }, { "epoch": 10.858123569794051, "grad_norm": 1.3252496719360352, "learning_rate": 4.01875e-07, "loss": 0.0046, "step": 23725 }, { "epoch": 10.869565217391305, "grad_norm": 0.7701499462127686, "learning_rate": 3.940625e-07, "loss": 0.0059, "step": 23750 }, { "epoch": 10.881006864988558, "grad_norm": 0.02328154630959034, "learning_rate": 3.8625e-07, "loss": 0.0063, "step": 23775 }, { "epoch": 10.892448512585812, "grad_norm": 1.4766663312911987, "learning_rate": 3.7843750000000003e-07, "loss": 0.0058, "step": 23800 }, { "epoch": 10.903890160183066, "grad_norm": 0.08133988082408905, "learning_rate": 3.70625e-07, "loss": 0.0032, "step": 23825 }, { "epoch": 10.915331807780321, "grad_norm": 9.198525428771973, "learning_rate": 3.628125e-07, "loss": 0.0054, "step": 23850 }, { "epoch": 10.926773455377575, "grad_norm": 0.09271799027919769, "learning_rate": 3.5500000000000004e-07, "loss": 0.0044, "step": 23875 }, { "epoch": 10.938215102974828, "grad_norm": 1.473221778869629, "learning_rate": 3.471875e-07, "loss": 0.0049, "step": 23900 }, { "epoch": 10.949656750572082, "grad_norm": 3.2264695167541504, "learning_rate": 3.3937500000000003e-07, "loss": 0.0049, "step": 23925 }, { "epoch": 10.961098398169337, "grad_norm": 14.144686698913574, "learning_rate": 3.315625e-07, "loss": 0.0072, "step": 23950 }, { "epoch": 10.972540045766591, "grad_norm": 2.536242723464966, "learning_rate": 3.2375e-07, "loss": 0.0047, "step": 23975 }, { "epoch": 10.983981693363845, "grad_norm": 7.181889533996582, "learning_rate": 3.159375e-07, "loss": 0.0123, "step": 24000 }, { "epoch": 10.995423340961098, "grad_norm": 0.28295642137527466, "learning_rate": 3.08125e-07, "loss": 0.005, "step": 24025 }, { "epoch": 11.006864988558352, "grad_norm": 1.317877173423767, "learning_rate": 3.0031250000000004e-07, "loss": 0.005, "step": 24050 }, { "epoch": 11.018306636155607, "grad_norm": 0.0772235319018364, "learning_rate": 2.925e-07, "loss": 0.0016, "step": 24075 }, { "epoch": 11.02974828375286, "grad_norm": 2.499826192855835, "learning_rate": 2.8468750000000003e-07, "loss": 0.0084, "step": 24100 }, { "epoch": 11.041189931350115, "grad_norm": 12.638914108276367, "learning_rate": 2.76875e-07, "loss": 0.004, "step": 24125 }, { "epoch": 11.052631578947368, "grad_norm": 0.1981966644525528, "learning_rate": 2.690625e-07, "loss": 0.0083, "step": 24150 }, { "epoch": 11.064073226544622, "grad_norm": 0.04735041409730911, "learning_rate": 2.6125e-07, "loss": 0.0054, "step": 24175 }, { "epoch": 11.075514874141877, "grad_norm": 0.5447256565093994, "learning_rate": 2.534375e-07, "loss": 0.0038, "step": 24200 }, { "epoch": 11.08695652173913, "grad_norm": 0.0347541943192482, "learning_rate": 2.4562500000000003e-07, "loss": 0.0029, "step": 24225 }, { "epoch": 11.098398169336384, "grad_norm": 13.768309593200684, "learning_rate": 2.3781250000000003e-07, "loss": 0.0049, "step": 24250 }, { "epoch": 11.109839816933638, "grad_norm": 1.426023244857788, "learning_rate": 2.3e-07, "loss": 0.0054, "step": 24275 }, { "epoch": 11.121281464530892, "grad_norm": 0.021928640082478523, "learning_rate": 2.221875e-07, "loss": 0.0063, "step": 24300 }, { "epoch": 11.132723112128147, "grad_norm": 2.05029296875, "learning_rate": 2.1437499999999999e-07, "loss": 0.0046, "step": 24325 }, { "epoch": 11.1441647597254, "grad_norm": 0.16184775531291962, "learning_rate": 2.0656250000000003e-07, "loss": 0.0062, "step": 24350 }, { "epoch": 11.155606407322654, "grad_norm": 1.0633113384246826, "learning_rate": 1.9875000000000003e-07, "loss": 0.0049, "step": 24375 }, { "epoch": 11.167048054919908, "grad_norm": 0.7184084057807922, "learning_rate": 1.9093750000000002e-07, "loss": 0.0088, "step": 24400 }, { "epoch": 11.178489702517162, "grad_norm": 0.5524642467498779, "learning_rate": 1.83125e-07, "loss": 0.0019, "step": 24425 }, { "epoch": 11.189931350114417, "grad_norm": 0.1881658285856247, "learning_rate": 1.753125e-07, "loss": 0.0061, "step": 24450 }, { "epoch": 11.20137299771167, "grad_norm": 3.7017602920532227, "learning_rate": 1.675e-07, "loss": 0.006, "step": 24475 }, { "epoch": 11.212814645308924, "grad_norm": 1.5420652627944946, "learning_rate": 1.596875e-07, "loss": 0.0048, "step": 24500 }, { "epoch": 11.224256292906178, "grad_norm": 5.843155384063721, "learning_rate": 1.51875e-07, "loss": 0.0064, "step": 24525 }, { "epoch": 11.235697940503432, "grad_norm": 0.06349411606788635, "learning_rate": 1.4406250000000002e-07, "loss": 0.0092, "step": 24550 }, { "epoch": 11.247139588100687, "grad_norm": 2.3772706985473633, "learning_rate": 1.3625000000000002e-07, "loss": 0.0029, "step": 24575 }, { "epoch": 11.25858123569794, "grad_norm": 1.1983476877212524, "learning_rate": 1.284375e-07, "loss": 0.0071, "step": 24600 }, { "epoch": 11.270022883295194, "grad_norm": 0.02243354730308056, "learning_rate": 1.20625e-07, "loss": 0.0035, "step": 24625 }, { "epoch": 11.281464530892448, "grad_norm": 1.8328884840011597, "learning_rate": 1.1312500000000002e-07, "loss": 0.0097, "step": 24650 }, { "epoch": 11.292906178489703, "grad_norm": 0.05551990494132042, "learning_rate": 1.0531250000000001e-07, "loss": 0.0034, "step": 24675 }, { "epoch": 11.304347826086957, "grad_norm": 0.030004534870386124, "learning_rate": 9.75e-08, "loss": 0.0103, "step": 24700 }, { "epoch": 11.31578947368421, "grad_norm": 0.03290963172912598, "learning_rate": 8.96875e-08, "loss": 0.0074, "step": 24725 }, { "epoch": 11.327231121281464, "grad_norm": 0.29857203364372253, "learning_rate": 8.187500000000001e-08, "loss": 0.0057, "step": 24750 }, { "epoch": 11.338672768878718, "grad_norm": 1.0099925994873047, "learning_rate": 7.40625e-08, "loss": 0.0066, "step": 24775 }, { "epoch": 11.350114416475973, "grad_norm": 0.26381850242614746, "learning_rate": 6.625e-08, "loss": 0.0063, "step": 24800 }, { "epoch": 11.361556064073227, "grad_norm": 0.07193097472190857, "learning_rate": 5.843750000000001e-08, "loss": 0.006, "step": 24825 }, { "epoch": 11.37299771167048, "grad_norm": 0.05023621395230293, "learning_rate": 5.0625e-08, "loss": 0.0032, "step": 24850 }, { "epoch": 11.384439359267734, "grad_norm": 0.03356996551156044, "learning_rate": 4.28125e-08, "loss": 0.0036, "step": 24875 }, { "epoch": 11.395881006864988, "grad_norm": 1.0069869756698608, "learning_rate": 3.5e-08, "loss": 0.0078, "step": 24900 }, { "epoch": 11.407322654462243, "grad_norm": 0.020418858155608177, "learning_rate": 2.7187499999999998e-08, "loss": 0.0024, "step": 24925 }, { "epoch": 11.418764302059497, "grad_norm": 0.03717822954058647, "learning_rate": 1.9375e-08, "loss": 0.0051, "step": 24950 }, { "epoch": 11.43020594965675, "grad_norm": 0.03731099143624306, "learning_rate": 1.1562500000000002e-08, "loss": 0.003, "step": 24975 }, { "epoch": 11.441647597254004, "grad_norm": 1.1690711975097656, "learning_rate": 3.75e-09, "loss": 0.0043, "step": 25000 }, { "epoch": 11.441647597254004, "eval_loss": 0.20993435382843018, "eval_runtime": 5536.2723, "eval_samples_per_second": 1.72, "eval_steps_per_second": 0.215, "eval_wer": 0.09534779548217828, "step": 25000 } ], "logging_steps": 25, "max_steps": 25000, "num_input_tokens_seen": 0, "num_train_epochs": 12, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.081970563920691e+20, "train_batch_size": 8, "trial_name": null, "trial_params": null }