{ "best_metric": null, "best_model_checkpoint": null, "epoch": 29.999559594239493, "eval_steps": 500, "global_step": 1021770, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01, "grad_norm": 1.2330610752105713, "learning_rate": 7.5e-05, "loss": 47.9921, "step": 500 }, { "epoch": 0.03, "grad_norm": 1.4777156114578247, "learning_rate": 0.00015, "loss": 6.0359, "step": 1000 }, { "epoch": 0.04, "grad_norm": 2.119438886642456, "learning_rate": 0.000225, "loss": 5.8858, "step": 1500 }, { "epoch": 0.06, "grad_norm": 2.3987252712249756, "learning_rate": 0.0003, "loss": 5.7995, "step": 2000 }, { "epoch": 0.07, "grad_norm": 1.466189980506897, "learning_rate": 0.0002998529080086686, "loss": 5.5973, "step": 2500 }, { "epoch": 0.09, "grad_norm": 1.6185613870620728, "learning_rate": 0.0002997058160173372, "loss": 4.5547, "step": 3000 }, { "epoch": 0.1, "grad_norm": 2.422328472137451, "learning_rate": 0.0002995587240260058, "loss": 3.8361, "step": 3500 }, { "epoch": 0.12, "grad_norm": 2.422811269760132, "learning_rate": 0.00029941163203467446, "loss": 3.5267, "step": 4000 }, { "epoch": 0.13, "grad_norm": 1.8635873794555664, "learning_rate": 0.0002992645400433431, "loss": 3.3318, "step": 4500 }, { "epoch": 0.15, "grad_norm": 2.5039353370666504, "learning_rate": 0.00029911744805201173, "loss": 3.2049, "step": 5000 }, { "epoch": 0.16, "grad_norm": 3.018263578414917, "learning_rate": 0.0002989703560606803, "loss": 3.1146, "step": 5500 }, { "epoch": 0.18, "grad_norm": 2.2091281414031982, "learning_rate": 0.00029882326406934894, "loss": 3.0412, "step": 6000 }, { "epoch": 0.19, "grad_norm": 2.7145626544952393, "learning_rate": 0.0002986761720780176, "loss": 2.9849, "step": 6500 }, { "epoch": 0.21, "grad_norm": 3.0529212951660156, "learning_rate": 0.00029852908008668616, "loss": 2.9731, "step": 7000 }, { "epoch": 0.22, "grad_norm": 2.4529542922973633, "learning_rate": 0.0002983819880953548, "loss": 2.9193, "step": 7500 }, { "epoch": 0.23, "grad_norm": 2.009725332260132, "learning_rate": 0.00029823489610402343, "loss": 2.8839, "step": 8000 }, { "epoch": 0.25, "grad_norm": 3.02736759185791, "learning_rate": 0.00029808780411269206, "loss": 2.8483, "step": 8500 }, { "epoch": 0.26, "grad_norm": 2.597303867340088, "learning_rate": 0.0002979407121213607, "loss": 2.84, "step": 9000 }, { "epoch": 0.28, "grad_norm": 2.826665163040161, "learning_rate": 0.0002977936201300293, "loss": 2.8151, "step": 9500 }, { "epoch": 0.29, "grad_norm": 2.141164779663086, "learning_rate": 0.0002976465281386979, "loss": 2.7756, "step": 10000 }, { "epoch": 0.31, "grad_norm": 3.0721323490142822, "learning_rate": 0.00029749943614736655, "loss": 2.7468, "step": 10500 }, { "epoch": 0.32, "grad_norm": 3.0305256843566895, "learning_rate": 0.0002973523441560351, "loss": 2.7548, "step": 11000 }, { "epoch": 0.34, "grad_norm": 1.7159794569015503, "learning_rate": 0.00029720525216470376, "loss": 2.739, "step": 11500 }, { "epoch": 0.35, "grad_norm": 4.015839099884033, "learning_rate": 0.0002970581601733724, "loss": 2.7338, "step": 12000 }, { "epoch": 0.37, "grad_norm": 2.1496126651763916, "learning_rate": 0.00029691106818204103, "loss": 2.7047, "step": 12500 }, { "epoch": 0.38, "grad_norm": 12.543700218200684, "learning_rate": 0.00029676397619070967, "loss": 2.6905, "step": 13000 }, { "epoch": 0.4, "grad_norm": 2.919098138809204, "learning_rate": 0.00029661688419937825, "loss": 2.6511, "step": 13500 }, { "epoch": 0.41, "grad_norm": 2.164848566055298, "learning_rate": 0.0002964697922080469, "loss": 2.6765, "step": 14000 }, { "epoch": 0.43, "grad_norm": 2.595182418823242, "learning_rate": 0.0002963227002167155, "loss": 2.6406, "step": 14500 }, { "epoch": 0.44, "grad_norm": 3.517666816711426, "learning_rate": 0.0002961756082253841, "loss": 2.6752, "step": 15000 }, { "epoch": 0.46, "grad_norm": 3.788487434387207, "learning_rate": 0.00029602851623405273, "loss": 2.6248, "step": 15500 }, { "epoch": 0.47, "grad_norm": 1.8919869661331177, "learning_rate": 0.00029588142424272137, "loss": 2.6751, "step": 16000 }, { "epoch": 0.48, "grad_norm": 7.1151628494262695, "learning_rate": 0.00029573433225139, "loss": 2.6024, "step": 16500 }, { "epoch": 0.5, "grad_norm": 1.6645219326019287, "learning_rate": 0.00029558724026005863, "loss": 2.6017, "step": 17000 }, { "epoch": 0.51, "grad_norm": 4.183793067932129, "learning_rate": 0.0002954401482687272, "loss": 2.5482, "step": 17500 }, { "epoch": 0.53, "grad_norm": 3.3193671703338623, "learning_rate": 0.00029529305627739585, "loss": 2.6058, "step": 18000 }, { "epoch": 0.54, "grad_norm": 1.861599326133728, "learning_rate": 0.0002951459642860645, "loss": 2.5766, "step": 18500 }, { "epoch": 0.56, "grad_norm": 2.443558931350708, "learning_rate": 0.00029499887229473306, "loss": 2.5901, "step": 19000 }, { "epoch": 0.57, "grad_norm": 2.1485493183135986, "learning_rate": 0.0002948517803034017, "loss": 2.549, "step": 19500 }, { "epoch": 0.59, "grad_norm": 2.802183151245117, "learning_rate": 0.00029470468831207033, "loss": 2.568, "step": 20000 }, { "epoch": 0.6, "grad_norm": 2.226308822631836, "learning_rate": 0.00029455759632073897, "loss": 2.5599, "step": 20500 }, { "epoch": 0.62, "grad_norm": 8.404243469238281, "learning_rate": 0.0002944105043294076, "loss": 2.5574, "step": 21000 }, { "epoch": 0.63, "grad_norm": 4.14946985244751, "learning_rate": 0.0002942634123380762, "loss": 2.5266, "step": 21500 }, { "epoch": 0.65, "grad_norm": 1.674615502357483, "learning_rate": 0.0002941163203467448, "loss": 2.5094, "step": 22000 }, { "epoch": 0.66, "grad_norm": 1.943082332611084, "learning_rate": 0.00029396922835541345, "loss": 2.4967, "step": 22500 }, { "epoch": 0.68, "grad_norm": 1.9026265144348145, "learning_rate": 0.0002938221363640821, "loss": 2.4974, "step": 23000 }, { "epoch": 0.69, "grad_norm": 4.214877128601074, "learning_rate": 0.00029367504437275067, "loss": 2.5329, "step": 23500 }, { "epoch": 0.7, "grad_norm": 2.1662731170654297, "learning_rate": 0.0002935279523814193, "loss": 2.508, "step": 24000 }, { "epoch": 0.72, "grad_norm": 2.6833200454711914, "learning_rate": 0.00029338086039008794, "loss": 2.4891, "step": 24500 }, { "epoch": 0.73, "grad_norm": 10.436029434204102, "learning_rate": 0.00029323376839875657, "loss": 2.506, "step": 25000 }, { "epoch": 0.75, "grad_norm": 1.6364326477050781, "learning_rate": 0.0002930866764074252, "loss": 2.5091, "step": 25500 }, { "epoch": 0.76, "grad_norm": 2.408642292022705, "learning_rate": 0.0002929395844160938, "loss": 2.4987, "step": 26000 }, { "epoch": 0.78, "grad_norm": 2.324516534805298, "learning_rate": 0.0002927924924247624, "loss": 2.5619, "step": 26500 }, { "epoch": 0.79, "grad_norm": 2.3488011360168457, "learning_rate": 0.00029264540043343106, "loss": 2.4594, "step": 27000 }, { "epoch": 0.81, "grad_norm": 2.130359411239624, "learning_rate": 0.00029249830844209964, "loss": 2.4677, "step": 27500 }, { "epoch": 0.82, "grad_norm": 3.6321048736572266, "learning_rate": 0.0002923512164507683, "loss": 2.4925, "step": 28000 }, { "epoch": 0.84, "grad_norm": 2.2566709518432617, "learning_rate": 0.0002922041244594369, "loss": 2.4779, "step": 28500 }, { "epoch": 0.85, "grad_norm": 1.4957294464111328, "learning_rate": 0.00029205703246810554, "loss": 2.4482, "step": 29000 }, { "epoch": 0.87, "grad_norm": 2.031252384185791, "learning_rate": 0.0002919099404767742, "loss": 2.4215, "step": 29500 }, { "epoch": 0.88, "grad_norm": 3.686828136444092, "learning_rate": 0.00029176284848544276, "loss": 2.4463, "step": 30000 }, { "epoch": 0.9, "grad_norm": 1.746907114982605, "learning_rate": 0.0002916157564941114, "loss": 2.4436, "step": 30500 }, { "epoch": 0.91, "grad_norm": 2.379333257675171, "learning_rate": 0.00029146866450278, "loss": 2.4244, "step": 31000 }, { "epoch": 0.92, "grad_norm": 2.696253776550293, "learning_rate": 0.0002913215725114486, "loss": 2.4218, "step": 31500 }, { "epoch": 0.94, "grad_norm": 4.1202826499938965, "learning_rate": 0.0002911744805201173, "loss": 2.438, "step": 32000 }, { "epoch": 0.95, "grad_norm": 3.3910372257232666, "learning_rate": 0.0002910273885287859, "loss": 2.4405, "step": 32500 }, { "epoch": 0.97, "grad_norm": 2.9638476371765137, "learning_rate": 0.0002908802965374545, "loss": 2.4461, "step": 33000 }, { "epoch": 0.98, "grad_norm": 3.280348062515259, "learning_rate": 0.00029073320454612315, "loss": 2.4177, "step": 33500 }, { "epoch": 1.0, "grad_norm": 2.654728651046753, "learning_rate": 0.0002905861125547917, "loss": 2.414, "step": 34000 }, { "epoch": 1.01, "grad_norm": 2.3180739879608154, "learning_rate": 0.00029043902056346036, "loss": 2.367, "step": 34500 }, { "epoch": 1.03, "grad_norm": 5.21975040435791, "learning_rate": 0.000290291928572129, "loss": 2.4158, "step": 35000 }, { "epoch": 1.04, "grad_norm": 1.7568458318710327, "learning_rate": 0.0002901448365807976, "loss": 2.3574, "step": 35500 }, { "epoch": 1.06, "grad_norm": 2.872269630432129, "learning_rate": 0.00028999774458946626, "loss": 2.3842, "step": 36000 }, { "epoch": 1.07, "grad_norm": 2.3413188457489014, "learning_rate": 0.00028985065259813484, "loss": 2.3313, "step": 36500 }, { "epoch": 1.09, "grad_norm": 1.4920777082443237, "learning_rate": 0.0002897035606068035, "loss": 2.4049, "step": 37000 }, { "epoch": 1.1, "grad_norm": 2.5813608169555664, "learning_rate": 0.0002895564686154721, "loss": 2.3641, "step": 37500 }, { "epoch": 1.12, "grad_norm": 2.110539674758911, "learning_rate": 0.0002894093766241407, "loss": 2.3704, "step": 38000 }, { "epoch": 1.13, "grad_norm": 2.397954225540161, "learning_rate": 0.00028926228463280933, "loss": 2.3962, "step": 38500 }, { "epoch": 1.15, "grad_norm": 1.290098786354065, "learning_rate": 0.00028911519264147796, "loss": 2.3359, "step": 39000 }, { "epoch": 1.16, "grad_norm": 1.7063418626785278, "learning_rate": 0.00028896810065014654, "loss": 2.34, "step": 39500 }, { "epoch": 1.17, "grad_norm": 1.603582501411438, "learning_rate": 0.00028882100865881523, "loss": 2.3439, "step": 40000 }, { "epoch": 1.19, "grad_norm": 1.6357001066207886, "learning_rate": 0.0002886739166674838, "loss": 2.3689, "step": 40500 }, { "epoch": 1.2, "grad_norm": 2.4969911575317383, "learning_rate": 0.00028852682467615245, "loss": 2.3621, "step": 41000 }, { "epoch": 1.22, "grad_norm": 1.7131437063217163, "learning_rate": 0.0002883797326848211, "loss": 2.3402, "step": 41500 }, { "epoch": 1.23, "grad_norm": 1.681779384613037, "learning_rate": 0.00028823264069348966, "loss": 2.3047, "step": 42000 }, { "epoch": 1.25, "grad_norm": 3.0767245292663574, "learning_rate": 0.0002880855487021583, "loss": 2.3529, "step": 42500 }, { "epoch": 1.26, "grad_norm": 2.4065330028533936, "learning_rate": 0.00028793845671082693, "loss": 2.3332, "step": 43000 }, { "epoch": 1.28, "grad_norm": 3.0359294414520264, "learning_rate": 0.0002877913647194955, "loss": 2.3477, "step": 43500 }, { "epoch": 1.29, "grad_norm": 2.010524272918701, "learning_rate": 0.0002876442727281642, "loss": 2.3123, "step": 44000 }, { "epoch": 1.31, "grad_norm": 9.476223945617676, "learning_rate": 0.0002874971807368328, "loss": 2.3187, "step": 44500 }, { "epoch": 1.32, "grad_norm": 1.5241541862487793, "learning_rate": 0.0002873500887455014, "loss": 2.3473, "step": 45000 }, { "epoch": 1.34, "grad_norm": 2.111661434173584, "learning_rate": 0.00028720299675417005, "loss": 2.3122, "step": 45500 }, { "epoch": 1.35, "grad_norm": 1.5378273725509644, "learning_rate": 0.00028705590476283863, "loss": 2.3185, "step": 46000 }, { "epoch": 1.37, "grad_norm": 2.218921184539795, "learning_rate": 0.00028690881277150727, "loss": 2.3426, "step": 46500 }, { "epoch": 1.38, "grad_norm": 2.234276294708252, "learning_rate": 0.0002867617207801759, "loss": 2.2885, "step": 47000 }, { "epoch": 1.39, "grad_norm": 2.929063081741333, "learning_rate": 0.0002866146287888445, "loss": 2.3047, "step": 47500 }, { "epoch": 1.41, "grad_norm": 1.764546513557434, "learning_rate": 0.00028646753679751317, "loss": 2.275, "step": 48000 }, { "epoch": 1.42, "grad_norm": 3.143376111984253, "learning_rate": 0.00028632044480618175, "loss": 2.2591, "step": 48500 }, { "epoch": 1.44, "grad_norm": 2.8672120571136475, "learning_rate": 0.0002861733528148504, "loss": 2.2462, "step": 49000 }, { "epoch": 1.45, "grad_norm": 3.577282667160034, "learning_rate": 0.000286026260823519, "loss": 2.2689, "step": 49500 }, { "epoch": 1.47, "grad_norm": 1.5024135112762451, "learning_rate": 0.0002858791688321876, "loss": 2.2886, "step": 50000 }, { "epoch": 1.48, "grad_norm": 2.36291241645813, "learning_rate": 0.00028573207684085624, "loss": 2.2102, "step": 50500 }, { "epoch": 1.5, "grad_norm": 2.7048864364624023, "learning_rate": 0.00028558498484952487, "loss": 2.2734, "step": 51000 }, { "epoch": 1.51, "grad_norm": 15.023186683654785, "learning_rate": 0.00028543789285819345, "loss": 2.2859, "step": 51500 }, { "epoch": 1.53, "grad_norm": 1.5235730409622192, "learning_rate": 0.00028529080086686214, "loss": 2.2459, "step": 52000 }, { "epoch": 1.54, "grad_norm": 3.157870292663574, "learning_rate": 0.0002851437088755307, "loss": 2.2679, "step": 52500 }, { "epoch": 1.56, "grad_norm": 5.78060245513916, "learning_rate": 0.00028499661688419936, "loss": 2.2781, "step": 53000 }, { "epoch": 1.57, "grad_norm": 3.3039705753326416, "learning_rate": 0.000284849524892868, "loss": 2.2761, "step": 53500 }, { "epoch": 1.59, "grad_norm": 2.286205530166626, "learning_rate": 0.00028470243290153657, "loss": 2.2549, "step": 54000 }, { "epoch": 1.6, "grad_norm": 2.2150983810424805, "learning_rate": 0.0002845553409102052, "loss": 2.2443, "step": 54500 }, { "epoch": 1.61, "grad_norm": 3.0205180644989014, "learning_rate": 0.00028440824891887384, "loss": 2.2676, "step": 55000 }, { "epoch": 1.63, "grad_norm": 1.5430960655212402, "learning_rate": 0.0002842611569275424, "loss": 2.2403, "step": 55500 }, { "epoch": 1.64, "grad_norm": 1.979278802871704, "learning_rate": 0.0002841140649362111, "loss": 2.2409, "step": 56000 }, { "epoch": 1.66, "grad_norm": 1.6833264827728271, "learning_rate": 0.0002839669729448797, "loss": 2.2657, "step": 56500 }, { "epoch": 1.67, "grad_norm": 1.9541752338409424, "learning_rate": 0.0002838198809535483, "loss": 2.2767, "step": 57000 }, { "epoch": 1.69, "grad_norm": 5.107935905456543, "learning_rate": 0.00028367278896221696, "loss": 2.2774, "step": 57500 }, { "epoch": 1.7, "grad_norm": 1.5604417324066162, "learning_rate": 0.00028352569697088554, "loss": 2.2405, "step": 58000 }, { "epoch": 1.72, "grad_norm": 1.6418412923812866, "learning_rate": 0.0002833786049795542, "loss": 2.2434, "step": 58500 }, { "epoch": 1.73, "grad_norm": 1.5278984308242798, "learning_rate": 0.0002832315129882228, "loss": 2.2977, "step": 59000 }, { "epoch": 1.75, "grad_norm": 2.3131840229034424, "learning_rate": 0.0002830844209968914, "loss": 2.2236, "step": 59500 }, { "epoch": 1.76, "grad_norm": 1.7341011762619019, "learning_rate": 0.0002829373290055601, "loss": 2.2235, "step": 60000 }, { "epoch": 1.78, "grad_norm": 2.8039567470550537, "learning_rate": 0.00028279023701422866, "loss": 2.2475, "step": 60500 }, { "epoch": 1.79, "grad_norm": 4.142283916473389, "learning_rate": 0.0002826431450228973, "loss": 2.2128, "step": 61000 }, { "epoch": 1.81, "grad_norm": 1.2194138765335083, "learning_rate": 0.00028249605303156593, "loss": 2.2568, "step": 61500 }, { "epoch": 1.82, "grad_norm": 1.2129665613174438, "learning_rate": 0.0002823489610402345, "loss": 2.2282, "step": 62000 }, { "epoch": 1.84, "grad_norm": 8.789237022399902, "learning_rate": 0.00028220186904890314, "loss": 2.2071, "step": 62500 }, { "epoch": 1.85, "grad_norm": 1.4489936828613281, "learning_rate": 0.0002820547770575718, "loss": 2.2389, "step": 63000 }, { "epoch": 1.86, "grad_norm": 1.591963768005371, "learning_rate": 0.0002819076850662404, "loss": 2.2277, "step": 63500 }, { "epoch": 1.88, "grad_norm": 2.642183303833008, "learning_rate": 0.00028176059307490905, "loss": 2.2195, "step": 64000 }, { "epoch": 1.89, "grad_norm": 2.7950453758239746, "learning_rate": 0.00028161350108357763, "loss": 2.2179, "step": 64500 }, { "epoch": 1.91, "grad_norm": 2.3858351707458496, "learning_rate": 0.00028146640909224626, "loss": 2.2266, "step": 65000 }, { "epoch": 1.92, "grad_norm": 2.2545394897460938, "learning_rate": 0.0002813193171009149, "loss": 2.2538, "step": 65500 }, { "epoch": 1.94, "grad_norm": 2.3457772731781006, "learning_rate": 0.0002811722251095835, "loss": 2.2351, "step": 66000 }, { "epoch": 1.95, "grad_norm": 2.0313475131988525, "learning_rate": 0.0002810251331182521, "loss": 2.2574, "step": 66500 }, { "epoch": 1.97, "grad_norm": 4.175018310546875, "learning_rate": 0.00028087804112692075, "loss": 2.2269, "step": 67000 }, { "epoch": 1.98, "grad_norm": 1.963953971862793, "learning_rate": 0.0002807309491355894, "loss": 2.2408, "step": 67500 }, { "epoch": 2.0, "grad_norm": 1.8645119667053223, "learning_rate": 0.000280583857144258, "loss": 2.2588, "step": 68000 }, { "epoch": 2.01, "grad_norm": 1.8845419883728027, "learning_rate": 0.0002804367651529266, "loss": 2.1532, "step": 68500 }, { "epoch": 2.03, "grad_norm": 2.3963561058044434, "learning_rate": 0.00028028967316159523, "loss": 2.1441, "step": 69000 }, { "epoch": 2.04, "grad_norm": 3.194061040878296, "learning_rate": 0.00028014258117026387, "loss": 2.1184, "step": 69500 }, { "epoch": 2.06, "grad_norm": 1.4685026407241821, "learning_rate": 0.0002799954891789325, "loss": 2.1504, "step": 70000 }, { "epoch": 2.07, "grad_norm": 6.733902931213379, "learning_rate": 0.0002798483971876011, "loss": 2.1449, "step": 70500 }, { "epoch": 2.08, "grad_norm": 1.5692353248596191, "learning_rate": 0.0002797013051962697, "loss": 2.1132, "step": 71000 }, { "epoch": 2.1, "grad_norm": 2.4298503398895264, "learning_rate": 0.00027955421320493835, "loss": 2.1557, "step": 71500 }, { "epoch": 2.11, "grad_norm": 1.9930598735809326, "learning_rate": 0.000279407121213607, "loss": 2.1509, "step": 72000 }, { "epoch": 2.13, "grad_norm": 3.2199394702911377, "learning_rate": 0.0002792600292222756, "loss": 2.1637, "step": 72500 }, { "epoch": 2.14, "grad_norm": 11.737617492675781, "learning_rate": 0.0002791129372309442, "loss": 2.1145, "step": 73000 }, { "epoch": 2.16, "grad_norm": 3.2281816005706787, "learning_rate": 0.00027896584523961284, "loss": 2.205, "step": 73500 }, { "epoch": 2.17, "grad_norm": 1.6318072080612183, "learning_rate": 0.00027881875324828147, "loss": 2.1355, "step": 74000 }, { "epoch": 2.19, "grad_norm": 1.4795607328414917, "learning_rate": 0.00027867166125695005, "loss": 2.1594, "step": 74500 }, { "epoch": 2.2, "grad_norm": 3.246556043624878, "learning_rate": 0.0002785245692656187, "loss": 2.1225, "step": 75000 }, { "epoch": 2.22, "grad_norm": 5.301700115203857, "learning_rate": 0.0002783774772742873, "loss": 2.1477, "step": 75500 }, { "epoch": 2.23, "grad_norm": 1.3937615156173706, "learning_rate": 0.00027823038528295595, "loss": 2.1377, "step": 76000 }, { "epoch": 2.25, "grad_norm": 5.8993611335754395, "learning_rate": 0.0002780832932916246, "loss": 2.156, "step": 76500 }, { "epoch": 2.26, "grad_norm": 3.7031867504119873, "learning_rate": 0.00027793620130029317, "loss": 2.1246, "step": 77000 }, { "epoch": 2.28, "grad_norm": 3.8393964767456055, "learning_rate": 0.0002777891093089618, "loss": 2.1503, "step": 77500 }, { "epoch": 2.29, "grad_norm": 1.3489915132522583, "learning_rate": 0.00027764201731763044, "loss": 2.1506, "step": 78000 }, { "epoch": 2.3, "grad_norm": 6.942753791809082, "learning_rate": 0.000277494925326299, "loss": 2.1575, "step": 78500 }, { "epoch": 2.32, "grad_norm": 3.2154929637908936, "learning_rate": 0.00027734783333496765, "loss": 2.1279, "step": 79000 }, { "epoch": 2.33, "grad_norm": 1.8168853521347046, "learning_rate": 0.0002772007413436363, "loss": 2.0715, "step": 79500 }, { "epoch": 2.35, "grad_norm": 2.112091541290283, "learning_rate": 0.0002770536493523049, "loss": 2.1344, "step": 80000 }, { "epoch": 2.36, "grad_norm": 2.050875663757324, "learning_rate": 0.00027690655736097356, "loss": 2.1416, "step": 80500 }, { "epoch": 2.38, "grad_norm": 1.736621618270874, "learning_rate": 0.00027675946536964214, "loss": 2.1196, "step": 81000 }, { "epoch": 2.39, "grad_norm": 2.051025629043579, "learning_rate": 0.0002766123733783108, "loss": 2.1272, "step": 81500 }, { "epoch": 2.41, "grad_norm": 4.314809799194336, "learning_rate": 0.0002764652813869794, "loss": 2.1618, "step": 82000 }, { "epoch": 2.42, "grad_norm": 2.920485496520996, "learning_rate": 0.000276318189395648, "loss": 2.1354, "step": 82500 }, { "epoch": 2.44, "grad_norm": 6.310970783233643, "learning_rate": 0.0002761710974043166, "loss": 2.1495, "step": 83000 }, { "epoch": 2.45, "grad_norm": 1.929152250289917, "learning_rate": 0.00027602400541298526, "loss": 2.1271, "step": 83500 }, { "epoch": 2.47, "grad_norm": 2.9410946369171143, "learning_rate": 0.0002758769134216539, "loss": 2.1646, "step": 84000 }, { "epoch": 2.48, "grad_norm": 2.474297523498535, "learning_rate": 0.00027572982143032253, "loss": 2.127, "step": 84500 }, { "epoch": 2.5, "grad_norm": 1.6442033052444458, "learning_rate": 0.0002755827294389911, "loss": 2.1433, "step": 85000 }, { "epoch": 2.51, "grad_norm": 1.3729546070098877, "learning_rate": 0.00027543563744765974, "loss": 2.1121, "step": 85500 }, { "epoch": 2.52, "grad_norm": 1.6497186422348022, "learning_rate": 0.0002752885454563284, "loss": 2.1566, "step": 86000 }, { "epoch": 2.54, "grad_norm": 2.6162164211273193, "learning_rate": 0.00027514145346499696, "loss": 2.1513, "step": 86500 }, { "epoch": 2.55, "grad_norm": 1.7324166297912598, "learning_rate": 0.0002749943614736656, "loss": 2.1482, "step": 87000 }, { "epoch": 2.57, "grad_norm": 1.3059158325195312, "learning_rate": 0.0002748472694823342, "loss": 2.1346, "step": 87500 }, { "epoch": 2.58, "grad_norm": 1.1019114255905151, "learning_rate": 0.00027470017749100286, "loss": 2.123, "step": 88000 }, { "epoch": 2.6, "grad_norm": 7.39063024520874, "learning_rate": 0.0002745530854996715, "loss": 2.1278, "step": 88500 }, { "epoch": 2.61, "grad_norm": 1.9237990379333496, "learning_rate": 0.0002744059935083401, "loss": 2.115, "step": 89000 }, { "epoch": 2.63, "grad_norm": 1.4632532596588135, "learning_rate": 0.0002742589015170087, "loss": 2.137, "step": 89500 }, { "epoch": 2.64, "grad_norm": 1.9587647914886475, "learning_rate": 0.00027411180952567735, "loss": 2.1056, "step": 90000 }, { "epoch": 2.66, "grad_norm": 7.492849349975586, "learning_rate": 0.0002739647175343459, "loss": 2.1208, "step": 90500 }, { "epoch": 2.67, "grad_norm": 1.7255750894546509, "learning_rate": 0.00027381762554301456, "loss": 2.1015, "step": 91000 }, { "epoch": 2.69, "grad_norm": 8.091690063476562, "learning_rate": 0.0002736705335516832, "loss": 2.1073, "step": 91500 }, { "epoch": 2.7, "grad_norm": 6.672662734985352, "learning_rate": 0.00027352344156035183, "loss": 2.1059, "step": 92000 }, { "epoch": 2.72, "grad_norm": 1.948398470878601, "learning_rate": 0.00027337634956902047, "loss": 2.0943, "step": 92500 }, { "epoch": 2.73, "grad_norm": 1.4260573387145996, "learning_rate": 0.00027322925757768905, "loss": 2.1283, "step": 93000 }, { "epoch": 2.75, "grad_norm": 1.5472320318222046, "learning_rate": 0.0002730821655863577, "loss": 2.0882, "step": 93500 }, { "epoch": 2.76, "grad_norm": 1.3860925436019897, "learning_rate": 0.0002729350735950263, "loss": 2.1323, "step": 94000 }, { "epoch": 2.77, "grad_norm": 2.232808828353882, "learning_rate": 0.0002727879816036949, "loss": 2.1485, "step": 94500 }, { "epoch": 2.79, "grad_norm": 1.8404427766799927, "learning_rate": 0.00027264088961236353, "loss": 2.1116, "step": 95000 }, { "epoch": 2.8, "grad_norm": 2.124656915664673, "learning_rate": 0.00027249379762103216, "loss": 2.1076, "step": 95500 }, { "epoch": 2.82, "grad_norm": 2.2169177532196045, "learning_rate": 0.0002723467056297008, "loss": 2.1035, "step": 96000 }, { "epoch": 2.83, "grad_norm": 2.7737512588500977, "learning_rate": 0.00027219961363836943, "loss": 2.1369, "step": 96500 }, { "epoch": 2.85, "grad_norm": 2.6662325859069824, "learning_rate": 0.000272052521647038, "loss": 2.1131, "step": 97000 }, { "epoch": 2.86, "grad_norm": 2.6663177013397217, "learning_rate": 0.00027190542965570665, "loss": 2.1175, "step": 97500 }, { "epoch": 2.88, "grad_norm": 1.446356177330017, "learning_rate": 0.0002717583376643753, "loss": 2.1076, "step": 98000 }, { "epoch": 2.89, "grad_norm": 2.3340353965759277, "learning_rate": 0.00027161124567304386, "loss": 2.0877, "step": 98500 }, { "epoch": 2.91, "grad_norm": 3.389127254486084, "learning_rate": 0.0002714641536817125, "loss": 2.0904, "step": 99000 }, { "epoch": 2.92, "grad_norm": 2.044728994369507, "learning_rate": 0.00027131706169038113, "loss": 2.0608, "step": 99500 }, { "epoch": 2.94, "grad_norm": 17.932655334472656, "learning_rate": 0.00027116996969904977, "loss": 2.1136, "step": 100000 }, { "epoch": 2.95, "grad_norm": 1.4239097833633423, "learning_rate": 0.0002710228777077184, "loss": 2.0783, "step": 100500 }, { "epoch": 2.97, "grad_norm": 2.1303598880767822, "learning_rate": 0.000270875785716387, "loss": 2.0981, "step": 101000 }, { "epoch": 2.98, "grad_norm": 2.2777888774871826, "learning_rate": 0.0002707286937250556, "loss": 2.0862, "step": 101500 }, { "epoch": 2.99, "grad_norm": 2.2035715579986572, "learning_rate": 0.00027058160173372425, "loss": 2.1089, "step": 102000 }, { "epoch": 3.01, "grad_norm": 2.7756547927856445, "learning_rate": 0.00027043450974239283, "loss": 2.0614, "step": 102500 }, { "epoch": 3.02, "grad_norm": 1.4222705364227295, "learning_rate": 0.00027028741775106147, "loss": 2.0216, "step": 103000 }, { "epoch": 3.04, "grad_norm": 1.5257524251937866, "learning_rate": 0.0002701403257597301, "loss": 2.0388, "step": 103500 }, { "epoch": 3.05, "grad_norm": 2.274021625518799, "learning_rate": 0.00026999323376839874, "loss": 2.048, "step": 104000 }, { "epoch": 3.07, "grad_norm": 5.903868198394775, "learning_rate": 0.00026984614177706737, "loss": 2.0008, "step": 104500 }, { "epoch": 3.08, "grad_norm": 1.561962604522705, "learning_rate": 0.00026969904978573595, "loss": 2.05, "step": 105000 }, { "epoch": 3.1, "grad_norm": 2.002523422241211, "learning_rate": 0.0002695519577944046, "loss": 2.0217, "step": 105500 }, { "epoch": 3.11, "grad_norm": 1.9345905780792236, "learning_rate": 0.0002694048658030732, "loss": 2.0343, "step": 106000 }, { "epoch": 3.13, "grad_norm": 1.5755488872528076, "learning_rate": 0.0002692577738117418, "loss": 2.0133, "step": 106500 }, { "epoch": 3.14, "grad_norm": 1.7304731607437134, "learning_rate": 0.00026911068182041044, "loss": 2.0177, "step": 107000 }, { "epoch": 3.16, "grad_norm": 2.4045250415802, "learning_rate": 0.00026896358982907907, "loss": 2.034, "step": 107500 }, { "epoch": 3.17, "grad_norm": 1.5433346033096313, "learning_rate": 0.0002688164978377477, "loss": 2.0228, "step": 108000 }, { "epoch": 3.19, "grad_norm": 2.033250093460083, "learning_rate": 0.00026866940584641634, "loss": 2.0076, "step": 108500 }, { "epoch": 3.2, "grad_norm": 3.0048422813415527, "learning_rate": 0.0002685223138550849, "loss": 1.9876, "step": 109000 }, { "epoch": 3.21, "grad_norm": 1.6355750560760498, "learning_rate": 0.00026837522186375356, "loss": 2.0259, "step": 109500 }, { "epoch": 3.23, "grad_norm": 7.954076290130615, "learning_rate": 0.0002682281298724222, "loss": 2.0408, "step": 110000 }, { "epoch": 3.24, "grad_norm": 1.3474091291427612, "learning_rate": 0.0002680810378810908, "loss": 2.0114, "step": 110500 }, { "epoch": 3.26, "grad_norm": 1.8665661811828613, "learning_rate": 0.0002679339458897594, "loss": 2.0388, "step": 111000 }, { "epoch": 3.27, "grad_norm": 1.452719807624817, "learning_rate": 0.00026778685389842804, "loss": 2.0411, "step": 111500 }, { "epoch": 3.29, "grad_norm": 1.9409444332122803, "learning_rate": 0.0002676397619070967, "loss": 2.0355, "step": 112000 }, { "epoch": 3.3, "grad_norm": 1.4139233827590942, "learning_rate": 0.0002674926699157653, "loss": 2.0306, "step": 112500 }, { "epoch": 3.32, "grad_norm": 2.548440933227539, "learning_rate": 0.00026734557792443394, "loss": 2.013, "step": 113000 }, { "epoch": 3.33, "grad_norm": 16.565500259399414, "learning_rate": 0.0002671984859331025, "loss": 2.0435, "step": 113500 }, { "epoch": 3.35, "grad_norm": 2.008643865585327, "learning_rate": 0.00026705139394177116, "loss": 2.0094, "step": 114000 }, { "epoch": 3.36, "grad_norm": 5.655598163604736, "learning_rate": 0.0002669043019504398, "loss": 2.0169, "step": 114500 }, { "epoch": 3.38, "grad_norm": 13.43127727508545, "learning_rate": 0.0002667572099591084, "loss": 2.0273, "step": 115000 }, { "epoch": 3.39, "grad_norm": 4.050173282623291, "learning_rate": 0.00026661011796777706, "loss": 2.0406, "step": 115500 }, { "epoch": 3.41, "grad_norm": 2.2970187664031982, "learning_rate": 0.00026646302597644564, "loss": 2.0374, "step": 116000 }, { "epoch": 3.42, "grad_norm": 1.8749233484268188, "learning_rate": 0.0002663159339851143, "loss": 1.991, "step": 116500 }, { "epoch": 3.44, "grad_norm": 2.4106335639953613, "learning_rate": 0.0002661688419937829, "loss": 2.0425, "step": 117000 }, { "epoch": 3.45, "grad_norm": 1.8036812543869019, "learning_rate": 0.0002660217500024515, "loss": 1.9968, "step": 117500 }, { "epoch": 3.46, "grad_norm": 1.831998348236084, "learning_rate": 0.00026587465801112013, "loss": 2.0291, "step": 118000 }, { "epoch": 3.48, "grad_norm": 2.4199304580688477, "learning_rate": 0.00026572756601978876, "loss": 2.0186, "step": 118500 }, { "epoch": 3.49, "grad_norm": 4.254742622375488, "learning_rate": 0.00026558047402845734, "loss": 2.0226, "step": 119000 }, { "epoch": 3.51, "grad_norm": 12.558996200561523, "learning_rate": 0.00026543338203712603, "loss": 2.0368, "step": 119500 }, { "epoch": 3.52, "grad_norm": 1.9092944860458374, "learning_rate": 0.0002652862900457946, "loss": 2.02, "step": 120000 }, { "epoch": 3.54, "grad_norm": 1.9921791553497314, "learning_rate": 0.00026513919805446325, "loss": 2.019, "step": 120500 }, { "epoch": 3.55, "grad_norm": 2.1361401081085205, "learning_rate": 0.0002649921060631319, "loss": 1.9927, "step": 121000 }, { "epoch": 3.57, "grad_norm": 1.6607388257980347, "learning_rate": 0.00026484501407180046, "loss": 2.0722, "step": 121500 }, { "epoch": 3.58, "grad_norm": 2.246613025665283, "learning_rate": 0.0002646979220804691, "loss": 2.0091, "step": 122000 }, { "epoch": 3.6, "grad_norm": 5.497014045715332, "learning_rate": 0.00026455083008913773, "loss": 2.0182, "step": 122500 }, { "epoch": 3.61, "grad_norm": 1.5741316080093384, "learning_rate": 0.00026440373809780637, "loss": 2.0113, "step": 123000 }, { "epoch": 3.63, "grad_norm": 2.1419429779052734, "learning_rate": 0.000264256646106475, "loss": 2.0112, "step": 123500 }, { "epoch": 3.64, "grad_norm": 2.309093952178955, "learning_rate": 0.0002641095541151436, "loss": 1.9943, "step": 124000 }, { "epoch": 3.66, "grad_norm": 1.7635866403579712, "learning_rate": 0.0002639624621238122, "loss": 2.0081, "step": 124500 }, { "epoch": 3.67, "grad_norm": 2.3035855293273926, "learning_rate": 0.00026381537013248085, "loss": 1.9658, "step": 125000 }, { "epoch": 3.68, "grad_norm": 1.8141978979110718, "learning_rate": 0.00026366827814114943, "loss": 2.0203, "step": 125500 }, { "epoch": 3.7, "grad_norm": 1.8287400007247925, "learning_rate": 0.00026352118614981807, "loss": 1.9636, "step": 126000 }, { "epoch": 3.71, "grad_norm": 1.356780767440796, "learning_rate": 0.0002633740941584867, "loss": 2.0116, "step": 126500 }, { "epoch": 3.73, "grad_norm": 5.973343849182129, "learning_rate": 0.00026322700216715534, "loss": 1.9837, "step": 127000 }, { "epoch": 3.74, "grad_norm": 2.090287446975708, "learning_rate": 0.00026307991017582397, "loss": 2.0161, "step": 127500 }, { "epoch": 3.76, "grad_norm": 1.422405481338501, "learning_rate": 0.00026293281818449255, "loss": 2.0096, "step": 128000 }, { "epoch": 3.77, "grad_norm": 5.0919365882873535, "learning_rate": 0.0002627857261931612, "loss": 1.997, "step": 128500 }, { "epoch": 3.79, "grad_norm": 2.7366178035736084, "learning_rate": 0.0002626386342018298, "loss": 2.001, "step": 129000 }, { "epoch": 3.8, "grad_norm": 1.3529456853866577, "learning_rate": 0.0002624915422104984, "loss": 2.0208, "step": 129500 }, { "epoch": 3.82, "grad_norm": 2.3386831283569336, "learning_rate": 0.00026234445021916704, "loss": 2.0049, "step": 130000 }, { "epoch": 3.83, "grad_norm": 2.074753999710083, "learning_rate": 0.00026219735822783567, "loss": 1.988, "step": 130500 }, { "epoch": 3.85, "grad_norm": 4.412100791931152, "learning_rate": 0.0002620502662365043, "loss": 1.9851, "step": 131000 }, { "epoch": 3.86, "grad_norm": 2.1146128177642822, "learning_rate": 0.00026190317424517294, "loss": 2.0049, "step": 131500 }, { "epoch": 3.88, "grad_norm": 7.746240615844727, "learning_rate": 0.0002617560822538415, "loss": 1.9978, "step": 132000 }, { "epoch": 3.89, "grad_norm": 1.8406715393066406, "learning_rate": 0.00026160899026251016, "loss": 1.9677, "step": 132500 }, { "epoch": 3.9, "grad_norm": 6.467206954956055, "learning_rate": 0.0002614618982711788, "loss": 1.9906, "step": 133000 }, { "epoch": 3.92, "grad_norm": 2.756176471710205, "learning_rate": 0.00026131480627984737, "loss": 2.0179, "step": 133500 }, { "epoch": 3.93, "grad_norm": 2.753931760787964, "learning_rate": 0.000261167714288516, "loss": 2.0166, "step": 134000 }, { "epoch": 3.95, "grad_norm": 32.68937301635742, "learning_rate": 0.00026102062229718464, "loss": 2.002, "step": 134500 }, { "epoch": 3.96, "grad_norm": 2.8605151176452637, "learning_rate": 0.0002608735303058533, "loss": 1.9997, "step": 135000 }, { "epoch": 3.98, "grad_norm": 1.3387537002563477, "learning_rate": 0.0002607264383145219, "loss": 1.9992, "step": 135500 }, { "epoch": 3.99, "grad_norm": 2.1721653938293457, "learning_rate": 0.0002605793463231905, "loss": 1.9946, "step": 136000 }, { "epoch": 4.01, "grad_norm": 15.696208953857422, "learning_rate": 0.0002604322543318591, "loss": 1.9808, "step": 136500 }, { "epoch": 4.02, "grad_norm": 2.118614912033081, "learning_rate": 0.00026028516234052776, "loss": 1.9334, "step": 137000 }, { "epoch": 4.04, "grad_norm": 4.258530616760254, "learning_rate": 0.00026013807034919634, "loss": 1.9505, "step": 137500 }, { "epoch": 4.05, "grad_norm": 1.8109312057495117, "learning_rate": 0.000259990978357865, "loss": 1.909, "step": 138000 }, { "epoch": 4.07, "grad_norm": 1.503454327583313, "learning_rate": 0.0002598438863665336, "loss": 1.9283, "step": 138500 }, { "epoch": 4.08, "grad_norm": 5.558799743652344, "learning_rate": 0.00025969679437520224, "loss": 1.959, "step": 139000 }, { "epoch": 4.1, "grad_norm": 2.509039878845215, "learning_rate": 0.0002595497023838709, "loss": 1.9111, "step": 139500 }, { "epoch": 4.11, "grad_norm": 1.6668161153793335, "learning_rate": 0.00025940261039253946, "loss": 1.9116, "step": 140000 }, { "epoch": 4.13, "grad_norm": 1.4484660625457764, "learning_rate": 0.0002592555184012081, "loss": 1.9337, "step": 140500 }, { "epoch": 4.14, "grad_norm": 2.352369785308838, "learning_rate": 0.00025910842640987673, "loss": 1.922, "step": 141000 }, { "epoch": 4.15, "grad_norm": 17.899917602539062, "learning_rate": 0.0002589613344185453, "loss": 1.9144, "step": 141500 }, { "epoch": 4.17, "grad_norm": 2.010423421859741, "learning_rate": 0.00025881424242721394, "loss": 1.9909, "step": 142000 }, { "epoch": 4.18, "grad_norm": 2.928180694580078, "learning_rate": 0.0002586671504358826, "loss": 1.9295, "step": 142500 }, { "epoch": 4.2, "grad_norm": 2.5269832611083984, "learning_rate": 0.0002585200584445512, "loss": 1.9473, "step": 143000 }, { "epoch": 4.21, "grad_norm": 4.112135410308838, "learning_rate": 0.00025837296645321985, "loss": 1.9538, "step": 143500 }, { "epoch": 4.23, "grad_norm": 2.3555727005004883, "learning_rate": 0.00025822587446188843, "loss": 1.951, "step": 144000 }, { "epoch": 4.24, "grad_norm": 2.098503828048706, "learning_rate": 0.00025807878247055706, "loss": 1.9492, "step": 144500 }, { "epoch": 4.26, "grad_norm": 2.459561824798584, "learning_rate": 0.0002579316904792257, "loss": 1.9427, "step": 145000 }, { "epoch": 4.27, "grad_norm": 2.0252935886383057, "learning_rate": 0.0002577845984878943, "loss": 1.9161, "step": 145500 }, { "epoch": 4.29, "grad_norm": 1.384768009185791, "learning_rate": 0.0002576375064965629, "loss": 1.9325, "step": 146000 }, { "epoch": 4.3, "grad_norm": 1.6305208206176758, "learning_rate": 0.00025749041450523155, "loss": 1.9892, "step": 146500 }, { "epoch": 4.32, "grad_norm": 2.4807193279266357, "learning_rate": 0.0002573433225139002, "loss": 1.9288, "step": 147000 }, { "epoch": 4.33, "grad_norm": 1.798276424407959, "learning_rate": 0.0002571962305225688, "loss": 1.9124, "step": 147500 }, { "epoch": 4.35, "grad_norm": 3.1262764930725098, "learning_rate": 0.0002570491385312374, "loss": 1.9356, "step": 148000 }, { "epoch": 4.36, "grad_norm": 2.5284624099731445, "learning_rate": 0.00025690204653990603, "loss": 1.9519, "step": 148500 }, { "epoch": 4.37, "grad_norm": 2.51084566116333, "learning_rate": 0.00025675495454857467, "loss": 1.948, "step": 149000 }, { "epoch": 4.39, "grad_norm": 1.7908506393432617, "learning_rate": 0.00025660786255724325, "loss": 1.9214, "step": 149500 }, { "epoch": 4.4, "grad_norm": 1.8897744417190552, "learning_rate": 0.0002564607705659119, "loss": 1.9492, "step": 150000 }, { "epoch": 4.42, "grad_norm": 2.1504998207092285, "learning_rate": 0.0002563136785745805, "loss": 1.9261, "step": 150500 }, { "epoch": 4.43, "grad_norm": 1.7902778387069702, "learning_rate": 0.00025616658658324915, "loss": 1.9332, "step": 151000 }, { "epoch": 4.45, "grad_norm": 1.6913944482803345, "learning_rate": 0.0002560194945919178, "loss": 1.9391, "step": 151500 }, { "epoch": 4.46, "grad_norm": 1.7668604850769043, "learning_rate": 0.00025587240260058637, "loss": 1.961, "step": 152000 }, { "epoch": 4.48, "grad_norm": 2.929547071456909, "learning_rate": 0.000255725310609255, "loss": 1.9578, "step": 152500 }, { "epoch": 4.49, "grad_norm": 2.5235610008239746, "learning_rate": 0.00025557821861792363, "loss": 1.8975, "step": 153000 }, { "epoch": 4.51, "grad_norm": 5.4155097007751465, "learning_rate": 0.0002554311266265922, "loss": 1.9549, "step": 153500 }, { "epoch": 4.52, "grad_norm": 2.9047696590423584, "learning_rate": 0.00025528403463526085, "loss": 1.9409, "step": 154000 }, { "epoch": 4.54, "grad_norm": 3.789259910583496, "learning_rate": 0.0002551369426439295, "loss": 1.9267, "step": 154500 }, { "epoch": 4.55, "grad_norm": 2.7584846019744873, "learning_rate": 0.0002549898506525981, "loss": 1.9138, "step": 155000 }, { "epoch": 4.57, "grad_norm": 4.33261251449585, "learning_rate": 0.00025484275866126675, "loss": 1.9611, "step": 155500 }, { "epoch": 4.58, "grad_norm": 3.2369813919067383, "learning_rate": 0.00025469566666993533, "loss": 1.9354, "step": 156000 }, { "epoch": 4.59, "grad_norm": 2.6031434535980225, "learning_rate": 0.00025454857467860397, "loss": 1.9172, "step": 156500 }, { "epoch": 4.61, "grad_norm": 2.9214913845062256, "learning_rate": 0.0002544014826872726, "loss": 1.9356, "step": 157000 }, { "epoch": 4.62, "grad_norm": 2.684860944747925, "learning_rate": 0.00025425439069594124, "loss": 1.9494, "step": 157500 }, { "epoch": 4.64, "grad_norm": 2.98620867729187, "learning_rate": 0.0002541072987046098, "loss": 1.9277, "step": 158000 }, { "epoch": 4.65, "grad_norm": 1.6060062646865845, "learning_rate": 0.00025396020671327845, "loss": 1.9417, "step": 158500 }, { "epoch": 4.67, "grad_norm": 1.4754972457885742, "learning_rate": 0.0002538131147219471, "loss": 1.9691, "step": 159000 }, { "epoch": 4.68, "grad_norm": 1.5829949378967285, "learning_rate": 0.0002536660227306157, "loss": 1.9201, "step": 159500 }, { "epoch": 4.7, "grad_norm": 2.13678240776062, "learning_rate": 0.00025351893073928436, "loss": 1.9697, "step": 160000 }, { "epoch": 4.71, "grad_norm": 2.030682325363159, "learning_rate": 0.00025337183874795294, "loss": 1.9364, "step": 160500 }, { "epoch": 4.73, "grad_norm": 2.715879201889038, "learning_rate": 0.00025322474675662157, "loss": 1.9574, "step": 161000 }, { "epoch": 4.74, "grad_norm": 48.37791442871094, "learning_rate": 0.0002530776547652902, "loss": 1.9372, "step": 161500 }, { "epoch": 4.76, "grad_norm": 3.232931613922119, "learning_rate": 0.0002529305627739588, "loss": 1.9093, "step": 162000 }, { "epoch": 4.77, "grad_norm": 2.2022180557250977, "learning_rate": 0.0002527834707826274, "loss": 1.8969, "step": 162500 }, { "epoch": 4.79, "grad_norm": 1.3633408546447754, "learning_rate": 0.00025263637879129606, "loss": 1.9476, "step": 163000 }, { "epoch": 4.8, "grad_norm": 1.8852393627166748, "learning_rate": 0.0002524892867999647, "loss": 1.9467, "step": 163500 }, { "epoch": 4.82, "grad_norm": 2.2378666400909424, "learning_rate": 0.0002523421948086333, "loss": 1.9216, "step": 164000 }, { "epoch": 4.83, "grad_norm": 6.068374156951904, "learning_rate": 0.0002521951028173019, "loss": 1.9188, "step": 164500 }, { "epoch": 4.84, "grad_norm": 2.9869394302368164, "learning_rate": 0.00025204801082597054, "loss": 1.8981, "step": 165000 }, { "epoch": 4.86, "grad_norm": 5.120233058929443, "learning_rate": 0.0002519009188346392, "loss": 1.9119, "step": 165500 }, { "epoch": 4.87, "grad_norm": 24.895925521850586, "learning_rate": 0.00025175382684330776, "loss": 1.9263, "step": 166000 }, { "epoch": 4.89, "grad_norm": 4.085075378417969, "learning_rate": 0.0002516067348519764, "loss": 1.9464, "step": 166500 }, { "epoch": 4.9, "grad_norm": 2.498352527618408, "learning_rate": 0.000251459642860645, "loss": 1.9349, "step": 167000 }, { "epoch": 4.92, "grad_norm": 1.9715830087661743, "learning_rate": 0.00025131255086931366, "loss": 1.9198, "step": 167500 }, { "epoch": 4.93, "grad_norm": 6.1134161949157715, "learning_rate": 0.0002511654588779823, "loss": 1.9064, "step": 168000 }, { "epoch": 4.95, "grad_norm": 1.4753895998001099, "learning_rate": 0.0002510183668866509, "loss": 1.9277, "step": 168500 }, { "epoch": 4.96, "grad_norm": 22.2166805267334, "learning_rate": 0.0002508712748953195, "loss": 1.9105, "step": 169000 }, { "epoch": 4.98, "grad_norm": 9.678267478942871, "learning_rate": 0.00025072418290398815, "loss": 1.9361, "step": 169500 }, { "epoch": 4.99, "grad_norm": 5.086581230163574, "learning_rate": 0.0002505770909126567, "loss": 1.9035, "step": 170000 }, { "epoch": 5.01, "grad_norm": 3.46391224861145, "learning_rate": 0.00025042999892132536, "loss": 1.9096, "step": 170500 }, { "epoch": 5.02, "grad_norm": 1.564864158630371, "learning_rate": 0.000250282906929994, "loss": 1.8412, "step": 171000 }, { "epoch": 5.04, "grad_norm": 3.917158603668213, "learning_rate": 0.00025013581493866263, "loss": 1.8263, "step": 171500 }, { "epoch": 5.05, "grad_norm": 2.3305134773254395, "learning_rate": 0.00024998872294733126, "loss": 1.8263, "step": 172000 }, { "epoch": 5.06, "grad_norm": 2.812856435775757, "learning_rate": 0.00024984163095599985, "loss": 1.8238, "step": 172500 }, { "epoch": 5.08, "grad_norm": 12.236431121826172, "learning_rate": 0.0002496945389646685, "loss": 1.8475, "step": 173000 }, { "epoch": 5.09, "grad_norm": 1.9508650302886963, "learning_rate": 0.0002495474469733371, "loss": 1.8671, "step": 173500 }, { "epoch": 5.11, "grad_norm": 4.95379638671875, "learning_rate": 0.0002494003549820057, "loss": 1.8541, "step": 174000 }, { "epoch": 5.12, "grad_norm": 2.8690032958984375, "learning_rate": 0.00024925326299067433, "loss": 1.8622, "step": 174500 }, { "epoch": 5.14, "grad_norm": 2.8076915740966797, "learning_rate": 0.00024910617099934296, "loss": 1.8394, "step": 175000 }, { "epoch": 5.15, "grad_norm": 1.6845248937606812, "learning_rate": 0.0002489590790080116, "loss": 1.8572, "step": 175500 }, { "epoch": 5.17, "grad_norm": 1.74410080909729, "learning_rate": 0.00024881198701668023, "loss": 1.8387, "step": 176000 }, { "epoch": 5.18, "grad_norm": 2.655266761779785, "learning_rate": 0.0002486648950253488, "loss": 1.8776, "step": 176500 }, { "epoch": 5.2, "grad_norm": 2.5884244441986084, "learning_rate": 0.00024851780303401745, "loss": 1.8564, "step": 177000 }, { "epoch": 5.21, "grad_norm": 4.314496040344238, "learning_rate": 0.0002483707110426861, "loss": 1.8573, "step": 177500 }, { "epoch": 5.23, "grad_norm": 2.15973162651062, "learning_rate": 0.00024822361905135466, "loss": 1.8671, "step": 178000 }, { "epoch": 5.24, "grad_norm": 1.7802814245224, "learning_rate": 0.00024807652706002335, "loss": 1.8722, "step": 178500 }, { "epoch": 5.26, "grad_norm": 29.04892921447754, "learning_rate": 0.00024792943506869193, "loss": 1.8747, "step": 179000 }, { "epoch": 5.27, "grad_norm": 2.0370380878448486, "learning_rate": 0.00024778234307736057, "loss": 1.8615, "step": 179500 }, { "epoch": 5.28, "grad_norm": 2.681140661239624, "learning_rate": 0.0002476352510860292, "loss": 1.8603, "step": 180000 }, { "epoch": 5.3, "grad_norm": 1.609506368637085, "learning_rate": 0.0002474881590946978, "loss": 1.857, "step": 180500 }, { "epoch": 5.31, "grad_norm": 1.4114452600479126, "learning_rate": 0.0002473410671033664, "loss": 1.8849, "step": 181000 }, { "epoch": 5.33, "grad_norm": 12.606436729431152, "learning_rate": 0.00024719397511203505, "loss": 1.8576, "step": 181500 }, { "epoch": 5.34, "grad_norm": 1.5284700393676758, "learning_rate": 0.00024704688312070363, "loss": 1.8565, "step": 182000 }, { "epoch": 5.36, "grad_norm": 2.134824514389038, "learning_rate": 0.0002468997911293723, "loss": 1.8447, "step": 182500 }, { "epoch": 5.37, "grad_norm": 2.755667209625244, "learning_rate": 0.0002467526991380409, "loss": 1.8894, "step": 183000 }, { "epoch": 5.39, "grad_norm": 1.8626573085784912, "learning_rate": 0.00024660560714670954, "loss": 1.8788, "step": 183500 }, { "epoch": 5.4, "grad_norm": 1.2436336278915405, "learning_rate": 0.00024645851515537817, "loss": 1.888, "step": 184000 }, { "epoch": 5.42, "grad_norm": 1.6450908184051514, "learning_rate": 0.00024631142316404675, "loss": 1.8576, "step": 184500 }, { "epoch": 5.43, "grad_norm": 2.5818614959716797, "learning_rate": 0.0002461643311727154, "loss": 1.8634, "step": 185000 }, { "epoch": 5.45, "grad_norm": 2.4842381477355957, "learning_rate": 0.000246017239181384, "loss": 1.8777, "step": 185500 }, { "epoch": 5.46, "grad_norm": 5.475472927093506, "learning_rate": 0.0002458701471900526, "loss": 1.8537, "step": 186000 }, { "epoch": 5.48, "grad_norm": 6.852906227111816, "learning_rate": 0.0002457230551987213, "loss": 1.863, "step": 186500 }, { "epoch": 5.49, "grad_norm": 2.8549883365631104, "learning_rate": 0.00024557596320738987, "loss": 1.8643, "step": 187000 }, { "epoch": 5.51, "grad_norm": 4.323022365570068, "learning_rate": 0.0002454288712160585, "loss": 1.8489, "step": 187500 }, { "epoch": 5.52, "grad_norm": 1.4015731811523438, "learning_rate": 0.00024528177922472714, "loss": 1.879, "step": 188000 }, { "epoch": 5.53, "grad_norm": 2.5605356693267822, "learning_rate": 0.0002451346872333957, "loss": 1.8564, "step": 188500 }, { "epoch": 5.55, "grad_norm": 1.567606806755066, "learning_rate": 0.00024498759524206436, "loss": 1.8902, "step": 189000 }, { "epoch": 5.56, "grad_norm": 1.4582362174987793, "learning_rate": 0.000244840503250733, "loss": 1.8659, "step": 189500 }, { "epoch": 5.58, "grad_norm": 1.8250012397766113, "learning_rate": 0.00024469341125940157, "loss": 1.8558, "step": 190000 }, { "epoch": 5.59, "grad_norm": 6.115236759185791, "learning_rate": 0.00024454631926807026, "loss": 1.839, "step": 190500 }, { "epoch": 5.61, "grad_norm": 4.505608081817627, "learning_rate": 0.00024439922727673884, "loss": 1.846, "step": 191000 }, { "epoch": 5.62, "grad_norm": 2.1621007919311523, "learning_rate": 0.0002442521352854075, "loss": 1.8287, "step": 191500 }, { "epoch": 5.64, "grad_norm": 2.337688684463501, "learning_rate": 0.0002441050432940761, "loss": 1.88, "step": 192000 }, { "epoch": 5.65, "grad_norm": 1.921249508857727, "learning_rate": 0.00024395795130274472, "loss": 1.8345, "step": 192500 }, { "epoch": 5.67, "grad_norm": 1.396031379699707, "learning_rate": 0.00024381085931141332, "loss": 1.8527, "step": 193000 }, { "epoch": 5.68, "grad_norm": 2.6552846431732178, "learning_rate": 0.00024366376732008196, "loss": 1.8628, "step": 193500 }, { "epoch": 5.7, "grad_norm": 4.992696285247803, "learning_rate": 0.00024351667532875057, "loss": 1.8727, "step": 194000 }, { "epoch": 5.71, "grad_norm": 4.128008842468262, "learning_rate": 0.00024336958333741923, "loss": 1.8677, "step": 194500 }, { "epoch": 5.73, "grad_norm": 2.2797539234161377, "learning_rate": 0.00024322249134608784, "loss": 1.8495, "step": 195000 }, { "epoch": 5.74, "grad_norm": 1.987032175064087, "learning_rate": 0.00024307539935475644, "loss": 1.8555, "step": 195500 }, { "epoch": 5.75, "grad_norm": 2.1700329780578613, "learning_rate": 0.00024292830736342508, "loss": 1.8745, "step": 196000 }, { "epoch": 5.77, "grad_norm": 2.0741491317749023, "learning_rate": 0.00024278121537209369, "loss": 1.8859, "step": 196500 }, { "epoch": 5.78, "grad_norm": 3.1270885467529297, "learning_rate": 0.0002426341233807623, "loss": 1.8778, "step": 197000 }, { "epoch": 5.8, "grad_norm": 2.7290618419647217, "learning_rate": 0.00024248703138943093, "loss": 1.8418, "step": 197500 }, { "epoch": 5.81, "grad_norm": 1.5647644996643066, "learning_rate": 0.00024233993939809954, "loss": 1.8519, "step": 198000 }, { "epoch": 5.83, "grad_norm": 2.4231462478637695, "learning_rate": 0.0002421928474067682, "loss": 1.868, "step": 198500 }, { "epoch": 5.84, "grad_norm": 3.1139655113220215, "learning_rate": 0.0002420457554154368, "loss": 1.8255, "step": 199000 }, { "epoch": 5.86, "grad_norm": 4.595509052276611, "learning_rate": 0.0002418986634241054, "loss": 1.8626, "step": 199500 }, { "epoch": 5.87, "grad_norm": 7.253945827484131, "learning_rate": 0.00024175157143277405, "loss": 1.8714, "step": 200000 }, { "epoch": 5.89, "grad_norm": 1.5077595710754395, "learning_rate": 0.00024160447944144266, "loss": 1.8743, "step": 200500 }, { "epoch": 5.9, "grad_norm": 5.791823387145996, "learning_rate": 0.00024145738745011126, "loss": 1.8499, "step": 201000 }, { "epoch": 5.92, "grad_norm": 2.558816909790039, "learning_rate": 0.0002413102954587799, "loss": 1.8757, "step": 201500 }, { "epoch": 5.93, "grad_norm": 2.4598610401153564, "learning_rate": 0.0002411632034674485, "loss": 1.8396, "step": 202000 }, { "epoch": 5.95, "grad_norm": 1.2680063247680664, "learning_rate": 0.00024101611147611717, "loss": 1.9265, "step": 202500 }, { "epoch": 5.96, "grad_norm": 13.615561485290527, "learning_rate": 0.00024086901948478577, "loss": 1.833, "step": 203000 }, { "epoch": 5.97, "grad_norm": 1.367218017578125, "learning_rate": 0.00024072192749345438, "loss": 1.836, "step": 203500 }, { "epoch": 5.99, "grad_norm": 1.8230597972869873, "learning_rate": 0.00024057483550212302, "loss": 1.8833, "step": 204000 }, { "epoch": 6.0, "grad_norm": 3.095012903213501, "learning_rate": 0.00024042774351079162, "loss": 1.8106, "step": 204500 }, { "epoch": 6.02, "grad_norm": 1.357537031173706, "learning_rate": 0.00024028065151946023, "loss": 1.809, "step": 205000 }, { "epoch": 6.03, "grad_norm": 1.8202615976333618, "learning_rate": 0.00024013355952812887, "loss": 1.788, "step": 205500 }, { "epoch": 6.05, "grad_norm": 2.0572142601013184, "learning_rate": 0.00023998646753679747, "loss": 1.7786, "step": 206000 }, { "epoch": 6.06, "grad_norm": 2.056692123413086, "learning_rate": 0.00023983937554546614, "loss": 1.801, "step": 206500 }, { "epoch": 6.08, "grad_norm": 1.6310349702835083, "learning_rate": 0.00023969228355413474, "loss": 1.794, "step": 207000 }, { "epoch": 6.09, "grad_norm": 2.9542150497436523, "learning_rate": 0.00023954519156280335, "loss": 1.8013, "step": 207500 }, { "epoch": 6.11, "grad_norm": 2.0560832023620605, "learning_rate": 0.00023939809957147199, "loss": 1.809, "step": 208000 }, { "epoch": 6.12, "grad_norm": 2.991140842437744, "learning_rate": 0.0002392510075801406, "loss": 1.8122, "step": 208500 }, { "epoch": 6.14, "grad_norm": 4.056896686553955, "learning_rate": 0.0002391039155888092, "loss": 1.7895, "step": 209000 }, { "epoch": 6.15, "grad_norm": 1.4278576374053955, "learning_rate": 0.00023895682359747784, "loss": 1.7778, "step": 209500 }, { "epoch": 6.17, "grad_norm": 2.918614387512207, "learning_rate": 0.00023880973160614644, "loss": 1.7882, "step": 210000 }, { "epoch": 6.18, "grad_norm": 6.002877712249756, "learning_rate": 0.0002386626396148151, "loss": 1.8129, "step": 210500 }, { "epoch": 6.2, "grad_norm": 1.8039432764053345, "learning_rate": 0.0002385155476234837, "loss": 1.7879, "step": 211000 }, { "epoch": 6.21, "grad_norm": 2.1559925079345703, "learning_rate": 0.00023836845563215232, "loss": 1.825, "step": 211500 }, { "epoch": 6.22, "grad_norm": 1.6617883443832397, "learning_rate": 0.00023822136364082095, "loss": 1.8048, "step": 212000 }, { "epoch": 6.24, "grad_norm": 3.7312443256378174, "learning_rate": 0.00023807427164948956, "loss": 1.7759, "step": 212500 }, { "epoch": 6.25, "grad_norm": 4.757487773895264, "learning_rate": 0.0002379271796581582, "loss": 1.8357, "step": 213000 }, { "epoch": 6.27, "grad_norm": 1.4285889863967896, "learning_rate": 0.0002377800876668268, "loss": 1.8127, "step": 213500 }, { "epoch": 6.28, "grad_norm": 1.7722175121307373, "learning_rate": 0.0002376329956754954, "loss": 1.79, "step": 214000 }, { "epoch": 6.3, "grad_norm": 2.17244291305542, "learning_rate": 0.00023748590368416407, "loss": 1.7949, "step": 214500 }, { "epoch": 6.31, "grad_norm": 2.136143922805786, "learning_rate": 0.00023733881169283268, "loss": 1.8199, "step": 215000 }, { "epoch": 6.33, "grad_norm": 1.36685049533844, "learning_rate": 0.00023719171970150132, "loss": 1.8101, "step": 215500 }, { "epoch": 6.34, "grad_norm": 1.5481481552124023, "learning_rate": 0.00023704462771016992, "loss": 1.819, "step": 216000 }, { "epoch": 6.36, "grad_norm": 2.1855366230010986, "learning_rate": 0.00023689753571883853, "loss": 1.8162, "step": 216500 }, { "epoch": 6.37, "grad_norm": 1.5019465684890747, "learning_rate": 0.00023675044372750717, "loss": 1.7929, "step": 217000 }, { "epoch": 6.39, "grad_norm": 1.4323623180389404, "learning_rate": 0.00023660335173617577, "loss": 1.7664, "step": 217500 }, { "epoch": 6.4, "grad_norm": 3.942918539047241, "learning_rate": 0.00023645625974484438, "loss": 1.8165, "step": 218000 }, { "epoch": 6.42, "grad_norm": 1.4521820545196533, "learning_rate": 0.00023630916775351304, "loss": 1.7966, "step": 218500 }, { "epoch": 6.43, "grad_norm": 3.9033968448638916, "learning_rate": 0.00023616207576218165, "loss": 1.7952, "step": 219000 }, { "epoch": 6.44, "grad_norm": 2.7818336486816406, "learning_rate": 0.00023601498377085029, "loss": 1.7963, "step": 219500 }, { "epoch": 6.46, "grad_norm": 4.406651496887207, "learning_rate": 0.0002358678917795189, "loss": 1.7901, "step": 220000 }, { "epoch": 6.47, "grad_norm": 2.1108040809631348, "learning_rate": 0.0002357207997881875, "loss": 1.7887, "step": 220500 }, { "epoch": 6.49, "grad_norm": 6.802318096160889, "learning_rate": 0.00023557370779685613, "loss": 1.7811, "step": 221000 }, { "epoch": 6.5, "grad_norm": 2.7774646282196045, "learning_rate": 0.00023542661580552474, "loss": 1.782, "step": 221500 }, { "epoch": 6.52, "grad_norm": 1.5817331075668335, "learning_rate": 0.00023527952381419335, "loss": 1.7883, "step": 222000 }, { "epoch": 6.53, "grad_norm": 1.3639189004898071, "learning_rate": 0.000235132431822862, "loss": 1.8137, "step": 222500 }, { "epoch": 6.55, "grad_norm": 3.018841505050659, "learning_rate": 0.00023498533983153062, "loss": 1.8012, "step": 223000 }, { "epoch": 6.56, "grad_norm": 1.6388349533081055, "learning_rate": 0.00023483824784019925, "loss": 1.7776, "step": 223500 }, { "epoch": 6.58, "grad_norm": 2.2828805446624756, "learning_rate": 0.00023469115584886786, "loss": 1.793, "step": 224000 }, { "epoch": 6.59, "grad_norm": 1.5646781921386719, "learning_rate": 0.00023454406385753647, "loss": 1.8049, "step": 224500 }, { "epoch": 6.61, "grad_norm": 2.5620057582855225, "learning_rate": 0.0002343969718662051, "loss": 1.8082, "step": 225000 }, { "epoch": 6.62, "grad_norm": 1.1866906881332397, "learning_rate": 0.0002342498798748737, "loss": 1.8197, "step": 225500 }, { "epoch": 6.64, "grad_norm": 8.060154914855957, "learning_rate": 0.00023410278788354232, "loss": 1.7765, "step": 226000 }, { "epoch": 6.65, "grad_norm": 2.2780892848968506, "learning_rate": 0.00023395569589221098, "loss": 1.7903, "step": 226500 }, { "epoch": 6.66, "grad_norm": 8.265579223632812, "learning_rate": 0.0002338086039008796, "loss": 1.8143, "step": 227000 }, { "epoch": 6.68, "grad_norm": 1.2793028354644775, "learning_rate": 0.00023366151190954822, "loss": 1.7729, "step": 227500 }, { "epoch": 6.69, "grad_norm": 1.7103880643844604, "learning_rate": 0.00023351441991821683, "loss": 1.8019, "step": 228000 }, { "epoch": 6.71, "grad_norm": 6.143857479095459, "learning_rate": 0.00023336732792688544, "loss": 1.7781, "step": 228500 }, { "epoch": 6.72, "grad_norm": 3.1208572387695312, "learning_rate": 0.00023322023593555407, "loss": 1.791, "step": 229000 }, { "epoch": 6.74, "grad_norm": 1.774464726448059, "learning_rate": 0.00023307314394422268, "loss": 1.7596, "step": 229500 }, { "epoch": 6.75, "grad_norm": 1.9884730577468872, "learning_rate": 0.0002329260519528913, "loss": 1.7842, "step": 230000 }, { "epoch": 6.77, "grad_norm": 0.9491617679595947, "learning_rate": 0.00023277895996155995, "loss": 1.8013, "step": 230500 }, { "epoch": 6.78, "grad_norm": 1.6505547761917114, "learning_rate": 0.00023263186797022856, "loss": 1.7913, "step": 231000 }, { "epoch": 6.8, "grad_norm": 1.262868046760559, "learning_rate": 0.0002324847759788972, "loss": 1.8212, "step": 231500 }, { "epoch": 6.81, "grad_norm": 1.1688213348388672, "learning_rate": 0.0002323376839875658, "loss": 1.815, "step": 232000 }, { "epoch": 6.83, "grad_norm": 1.4348254203796387, "learning_rate": 0.0002321905919962344, "loss": 1.8335, "step": 232500 }, { "epoch": 6.84, "grad_norm": 1.977734923362732, "learning_rate": 0.00023204350000490304, "loss": 1.8231, "step": 233000 }, { "epoch": 6.86, "grad_norm": 12.462953567504883, "learning_rate": 0.00023189640801357165, "loss": 1.8086, "step": 233500 }, { "epoch": 6.87, "grad_norm": 1.9486029148101807, "learning_rate": 0.00023174931602224028, "loss": 1.7932, "step": 234000 }, { "epoch": 6.89, "grad_norm": 6.036319732666016, "learning_rate": 0.00023160222403090892, "loss": 1.7906, "step": 234500 }, { "epoch": 6.9, "grad_norm": 1.6775847673416138, "learning_rate": 0.00023145513203957753, "loss": 1.8026, "step": 235000 }, { "epoch": 6.91, "grad_norm": 2.1960055828094482, "learning_rate": 0.00023130804004824616, "loss": 1.8093, "step": 235500 }, { "epoch": 6.93, "grad_norm": 4.2877197265625, "learning_rate": 0.00023116094805691477, "loss": 1.8198, "step": 236000 }, { "epoch": 6.94, "grad_norm": 1.4908256530761719, "learning_rate": 0.0002310138560655834, "loss": 1.7882, "step": 236500 }, { "epoch": 6.96, "grad_norm": 5.148675918579102, "learning_rate": 0.000230866764074252, "loss": 1.786, "step": 237000 }, { "epoch": 6.97, "grad_norm": 5.5215744972229, "learning_rate": 0.00023071967208292062, "loss": 1.8168, "step": 237500 }, { "epoch": 6.99, "grad_norm": 21.99406623840332, "learning_rate": 0.00023057258009158928, "loss": 1.7993, "step": 238000 }, { "epoch": 7.0, "grad_norm": 3.8185484409332275, "learning_rate": 0.0002304254881002579, "loss": 1.8105, "step": 238500 }, { "epoch": 7.02, "grad_norm": 7.100237846374512, "learning_rate": 0.00023027839610892652, "loss": 1.7102, "step": 239000 }, { "epoch": 7.03, "grad_norm": 5.007193088531494, "learning_rate": 0.00023013130411759513, "loss": 1.7293, "step": 239500 }, { "epoch": 7.05, "grad_norm": 1.6522769927978516, "learning_rate": 0.00022998421212626374, "loss": 1.7199, "step": 240000 }, { "epoch": 7.06, "grad_norm": 1.7635401487350464, "learning_rate": 0.00022983712013493237, "loss": 1.7304, "step": 240500 }, { "epoch": 7.08, "grad_norm": 2.441962957382202, "learning_rate": 0.00022969002814360098, "loss": 1.7291, "step": 241000 }, { "epoch": 7.09, "grad_norm": 1.6155414581298828, "learning_rate": 0.0002295429361522696, "loss": 1.7477, "step": 241500 }, { "epoch": 7.11, "grad_norm": 1.3782049417495728, "learning_rate": 0.00022939584416093825, "loss": 1.7552, "step": 242000 }, { "epoch": 7.12, "grad_norm": 1.6964011192321777, "learning_rate": 0.00022924875216960686, "loss": 1.7233, "step": 242500 }, { "epoch": 7.13, "grad_norm": 1.9475128650665283, "learning_rate": 0.0002291016601782755, "loss": 1.7603, "step": 243000 }, { "epoch": 7.15, "grad_norm": 2.674274444580078, "learning_rate": 0.0002289545681869441, "loss": 1.7472, "step": 243500 }, { "epoch": 7.16, "grad_norm": 1.6163575649261475, "learning_rate": 0.0002288074761956127, "loss": 1.7754, "step": 244000 }, { "epoch": 7.18, "grad_norm": 1.3212227821350098, "learning_rate": 0.00022866038420428134, "loss": 1.7828, "step": 244500 }, { "epoch": 7.19, "grad_norm": 1.9953806400299072, "learning_rate": 0.00022851329221294995, "loss": 1.7289, "step": 245000 }, { "epoch": 7.21, "grad_norm": 1.5477960109710693, "learning_rate": 0.00022836620022161856, "loss": 1.7651, "step": 245500 }, { "epoch": 7.22, "grad_norm": 16.499107360839844, "learning_rate": 0.00022821910823028722, "loss": 1.7207, "step": 246000 }, { "epoch": 7.24, "grad_norm": 2.4671411514282227, "learning_rate": 0.00022807201623895583, "loss": 1.7438, "step": 246500 }, { "epoch": 7.25, "grad_norm": 4.774916648864746, "learning_rate": 0.00022792492424762446, "loss": 1.7721, "step": 247000 }, { "epoch": 7.27, "grad_norm": 2.282515287399292, "learning_rate": 0.00022777783225629307, "loss": 1.7145, "step": 247500 }, { "epoch": 7.28, "grad_norm": 1.9206039905548096, "learning_rate": 0.00022763074026496168, "loss": 1.7314, "step": 248000 }, { "epoch": 7.3, "grad_norm": 2.7325820922851562, "learning_rate": 0.0002274836482736303, "loss": 1.725, "step": 248500 }, { "epoch": 7.31, "grad_norm": 1.7279201745986938, "learning_rate": 0.00022733655628229892, "loss": 1.7095, "step": 249000 }, { "epoch": 7.33, "grad_norm": 1.8310534954071045, "learning_rate": 0.00022718946429096753, "loss": 1.7325, "step": 249500 }, { "epoch": 7.34, "grad_norm": 5.919315338134766, "learning_rate": 0.0002270423722996362, "loss": 1.7301, "step": 250000 }, { "epoch": 7.35, "grad_norm": 2.3758251667022705, "learning_rate": 0.0002268952803083048, "loss": 1.7529, "step": 250500 }, { "epoch": 7.37, "grad_norm": 3.524482011795044, "learning_rate": 0.00022674818831697343, "loss": 1.7822, "step": 251000 }, { "epoch": 7.38, "grad_norm": 2.1281306743621826, "learning_rate": 0.00022660109632564204, "loss": 1.7315, "step": 251500 }, { "epoch": 7.4, "grad_norm": 1.6832419633865356, "learning_rate": 0.00022645400433431064, "loss": 1.7553, "step": 252000 }, { "epoch": 7.41, "grad_norm": 13.24813175201416, "learning_rate": 0.00022630691234297928, "loss": 1.7623, "step": 252500 }, { "epoch": 7.43, "grad_norm": 1.5083931684494019, "learning_rate": 0.0002261598203516479, "loss": 1.7579, "step": 253000 }, { "epoch": 7.44, "grad_norm": 3.207942247390747, "learning_rate": 0.00022601272836031652, "loss": 1.7248, "step": 253500 }, { "epoch": 7.46, "grad_norm": 7.897548675537109, "learning_rate": 0.00022586563636898516, "loss": 1.7309, "step": 254000 }, { "epoch": 7.47, "grad_norm": 2.2789628505706787, "learning_rate": 0.00022571854437765376, "loss": 1.7444, "step": 254500 }, { "epoch": 7.49, "grad_norm": 2.129913568496704, "learning_rate": 0.0002255714523863224, "loss": 1.7505, "step": 255000 }, { "epoch": 7.5, "grad_norm": 2.0121564865112305, "learning_rate": 0.000225424360394991, "loss": 1.8085, "step": 255500 }, { "epoch": 7.52, "grad_norm": 3.321162462234497, "learning_rate": 0.00022527726840365964, "loss": 1.7601, "step": 256000 }, { "epoch": 7.53, "grad_norm": 8.002218246459961, "learning_rate": 0.00022513017641232825, "loss": 1.7719, "step": 256500 }, { "epoch": 7.55, "grad_norm": 1.610587477684021, "learning_rate": 0.00022498308442099686, "loss": 1.7373, "step": 257000 }, { "epoch": 7.56, "grad_norm": 5.272850513458252, "learning_rate": 0.0002248359924296655, "loss": 1.7481, "step": 257500 }, { "epoch": 7.57, "grad_norm": 1.519241452217102, "learning_rate": 0.00022468890043833413, "loss": 1.7722, "step": 258000 }, { "epoch": 7.59, "grad_norm": 3.9072492122650146, "learning_rate": 0.00022454180844700276, "loss": 1.7425, "step": 258500 }, { "epoch": 7.6, "grad_norm": 16.145322799682617, "learning_rate": 0.00022439471645567137, "loss": 1.7645, "step": 259000 }, { "epoch": 7.62, "grad_norm": 1.7309913635253906, "learning_rate": 0.00022424762446433998, "loss": 1.7388, "step": 259500 }, { "epoch": 7.63, "grad_norm": 1.4854837656021118, "learning_rate": 0.0002241005324730086, "loss": 1.7728, "step": 260000 }, { "epoch": 7.65, "grad_norm": 1.9065077304840088, "learning_rate": 0.00022395344048167722, "loss": 1.7306, "step": 260500 }, { "epoch": 7.66, "grad_norm": 2.6869423389434814, "learning_rate": 0.00022380634849034582, "loss": 1.751, "step": 261000 }, { "epoch": 7.68, "grad_norm": 2.3300232887268066, "learning_rate": 0.00022365925649901446, "loss": 1.7643, "step": 261500 }, { "epoch": 7.69, "grad_norm": 1.6109992265701294, "learning_rate": 0.0002235121645076831, "loss": 1.7848, "step": 262000 }, { "epoch": 7.71, "grad_norm": 3.2768895626068115, "learning_rate": 0.00022336507251635173, "loss": 1.7638, "step": 262500 }, { "epoch": 7.72, "grad_norm": 4.605926036834717, "learning_rate": 0.00022321798052502034, "loss": 1.7801, "step": 263000 }, { "epoch": 7.74, "grad_norm": 1.451865553855896, "learning_rate": 0.00022307088853368894, "loss": 1.7229, "step": 263500 }, { "epoch": 7.75, "grad_norm": 1.938750982284546, "learning_rate": 0.00022292379654235758, "loss": 1.7575, "step": 264000 }, { "epoch": 7.77, "grad_norm": 8.690546989440918, "learning_rate": 0.00022277670455102619, "loss": 1.7572, "step": 264500 }, { "epoch": 7.78, "grad_norm": 6.99992036819458, "learning_rate": 0.0002226296125596948, "loss": 1.7912, "step": 265000 }, { "epoch": 7.8, "grad_norm": 1.375183343887329, "learning_rate": 0.00022248252056836343, "loss": 1.7588, "step": 265500 }, { "epoch": 7.81, "grad_norm": 1.4782127141952515, "learning_rate": 0.00022233542857703206, "loss": 1.7929, "step": 266000 }, { "epoch": 7.82, "grad_norm": 3.282301187515259, "learning_rate": 0.0002221883365857007, "loss": 1.7855, "step": 266500 }, { "epoch": 7.84, "grad_norm": 2.2392537593841553, "learning_rate": 0.0002220412445943693, "loss": 1.7603, "step": 267000 }, { "epoch": 7.85, "grad_norm": 3.2060225009918213, "learning_rate": 0.0002218941526030379, "loss": 1.7416, "step": 267500 }, { "epoch": 7.87, "grad_norm": 2.6562771797180176, "learning_rate": 0.00022174706061170655, "loss": 1.766, "step": 268000 }, { "epoch": 7.88, "grad_norm": 3.6191391944885254, "learning_rate": 0.00022159996862037516, "loss": 1.7746, "step": 268500 }, { "epoch": 7.9, "grad_norm": 1.9865895509719849, "learning_rate": 0.00022145287662904376, "loss": 1.7841, "step": 269000 }, { "epoch": 7.91, "grad_norm": 1.3776191473007202, "learning_rate": 0.0002213057846377124, "loss": 1.7709, "step": 269500 }, { "epoch": 7.93, "grad_norm": 2.238267183303833, "learning_rate": 0.00022115869264638103, "loss": 1.7764, "step": 270000 }, { "epoch": 7.94, "grad_norm": 3.3425824642181396, "learning_rate": 0.00022101160065504967, "loss": 1.7901, "step": 270500 }, { "epoch": 7.96, "grad_norm": 2.183436632156372, "learning_rate": 0.00022086450866371827, "loss": 1.757, "step": 271000 }, { "epoch": 7.97, "grad_norm": 3.393548011779785, "learning_rate": 0.00022071741667238688, "loss": 1.7538, "step": 271500 }, { "epoch": 7.99, "grad_norm": 4.727015018463135, "learning_rate": 0.00022057032468105552, "loss": 1.7585, "step": 272000 }, { "epoch": 8.0, "grad_norm": 2.0307464599609375, "learning_rate": 0.00022042323268972412, "loss": 1.7659, "step": 272500 }, { "epoch": 8.02, "grad_norm": 4.098311424255371, "learning_rate": 0.00022027614069839273, "loss": 1.7191, "step": 273000 }, { "epoch": 8.03, "grad_norm": 2.290457248687744, "learning_rate": 0.00022012904870706137, "loss": 1.7206, "step": 273500 }, { "epoch": 8.04, "grad_norm": 2.1348979473114014, "learning_rate": 0.00021998195671573, "loss": 1.6935, "step": 274000 }, { "epoch": 8.06, "grad_norm": 2.5333049297332764, "learning_rate": 0.00021983486472439864, "loss": 1.6709, "step": 274500 }, { "epoch": 8.07, "grad_norm": 2.1333959102630615, "learning_rate": 0.00021968777273306724, "loss": 1.7318, "step": 275000 }, { "epoch": 8.09, "grad_norm": 1.6256059408187866, "learning_rate": 0.00021954068074173585, "loss": 1.7007, "step": 275500 }, { "epoch": 8.1, "grad_norm": 1.4450695514678955, "learning_rate": 0.00021939358875040449, "loss": 1.7038, "step": 276000 }, { "epoch": 8.12, "grad_norm": 4.230172634124756, "learning_rate": 0.0002192464967590731, "loss": 1.6738, "step": 276500 }, { "epoch": 8.13, "grad_norm": 2.046879529953003, "learning_rate": 0.00021909940476774173, "loss": 1.6658, "step": 277000 }, { "epoch": 8.15, "grad_norm": 2.2633895874023438, "learning_rate": 0.00021895231277641034, "loss": 1.6929, "step": 277500 }, { "epoch": 8.16, "grad_norm": 34.2008171081543, "learning_rate": 0.00021880522078507897, "loss": 1.6956, "step": 278000 }, { "epoch": 8.18, "grad_norm": 11.11546516418457, "learning_rate": 0.0002186581287937476, "loss": 1.7103, "step": 278500 }, { "epoch": 8.19, "grad_norm": 5.735045909881592, "learning_rate": 0.0002185110368024162, "loss": 1.6991, "step": 279000 }, { "epoch": 8.21, "grad_norm": 2.9397599697113037, "learning_rate": 0.00021836394481108485, "loss": 1.6909, "step": 279500 }, { "epoch": 8.22, "grad_norm": 1.9226768016815186, "learning_rate": 0.00021821685281975345, "loss": 1.6894, "step": 280000 }, { "epoch": 8.24, "grad_norm": 2.0962398052215576, "learning_rate": 0.00021806976082842206, "loss": 1.7094, "step": 280500 }, { "epoch": 8.25, "grad_norm": 13.345786094665527, "learning_rate": 0.0002179226688370907, "loss": 1.6927, "step": 281000 }, { "epoch": 8.26, "grad_norm": 3.1707746982574463, "learning_rate": 0.0002177755768457593, "loss": 1.7004, "step": 281500 }, { "epoch": 8.28, "grad_norm": 3.3897292613983154, "learning_rate": 0.00021762848485442797, "loss": 1.7094, "step": 282000 }, { "epoch": 8.29, "grad_norm": 5.791382312774658, "learning_rate": 0.00021748139286309657, "loss": 1.7279, "step": 282500 }, { "epoch": 8.31, "grad_norm": 2.3419294357299805, "learning_rate": 0.00021733430087176518, "loss": 1.7053, "step": 283000 }, { "epoch": 8.32, "grad_norm": 1.9053447246551514, "learning_rate": 0.00021718720888043382, "loss": 1.7225, "step": 283500 }, { "epoch": 8.34, "grad_norm": 1.8822693824768066, "learning_rate": 0.00021704011688910242, "loss": 1.6968, "step": 284000 }, { "epoch": 8.35, "grad_norm": 1.4814122915267944, "learning_rate": 0.00021689302489777103, "loss": 1.6697, "step": 284500 }, { "epoch": 8.37, "grad_norm": 8.579392433166504, "learning_rate": 0.00021674593290643967, "loss": 1.7206, "step": 285000 }, { "epoch": 8.38, "grad_norm": 1.9921379089355469, "learning_rate": 0.00021659884091510827, "loss": 1.7113, "step": 285500 }, { "epoch": 8.4, "grad_norm": 2.2425730228424072, "learning_rate": 0.00021645174892377694, "loss": 1.6893, "step": 286000 }, { "epoch": 8.41, "grad_norm": 4.465683460235596, "learning_rate": 0.00021630465693244554, "loss": 1.6997, "step": 286500 }, { "epoch": 8.43, "grad_norm": 11.377062797546387, "learning_rate": 0.00021615756494111415, "loss": 1.691, "step": 287000 }, { "epoch": 8.44, "grad_norm": 4.9282636642456055, "learning_rate": 0.00021601047294978279, "loss": 1.6722, "step": 287500 }, { "epoch": 8.46, "grad_norm": 4.008066177368164, "learning_rate": 0.0002158633809584514, "loss": 1.6498, "step": 288000 }, { "epoch": 8.47, "grad_norm": 1.7437248229980469, "learning_rate": 0.00021571628896712, "loss": 1.7336, "step": 288500 }, { "epoch": 8.49, "grad_norm": 1.627820372581482, "learning_rate": 0.00021556919697578864, "loss": 1.6973, "step": 289000 }, { "epoch": 8.5, "grad_norm": 5.1567254066467285, "learning_rate": 0.00021542210498445724, "loss": 1.6974, "step": 289500 }, { "epoch": 8.51, "grad_norm": 1.3468036651611328, "learning_rate": 0.0002152750129931259, "loss": 1.6815, "step": 290000 }, { "epoch": 8.53, "grad_norm": 1.6024253368377686, "learning_rate": 0.0002151279210017945, "loss": 1.6784, "step": 290500 }, { "epoch": 8.54, "grad_norm": 1.915425181388855, "learning_rate": 0.00021498082901046312, "loss": 1.7297, "step": 291000 }, { "epoch": 8.56, "grad_norm": 2.750544548034668, "learning_rate": 0.00021483373701913175, "loss": 1.717, "step": 291500 }, { "epoch": 8.57, "grad_norm": 8.173510551452637, "learning_rate": 0.00021468664502780036, "loss": 1.6971, "step": 292000 }, { "epoch": 8.59, "grad_norm": 2.2209818363189697, "learning_rate": 0.00021453955303646897, "loss": 1.6803, "step": 292500 }, { "epoch": 8.6, "grad_norm": 4.283031463623047, "learning_rate": 0.0002143924610451376, "loss": 1.7101, "step": 293000 }, { "epoch": 8.62, "grad_norm": 2.29679799079895, "learning_rate": 0.0002142453690538062, "loss": 1.7021, "step": 293500 }, { "epoch": 8.63, "grad_norm": 3.301649570465088, "learning_rate": 0.00021409827706247487, "loss": 1.727, "step": 294000 }, { "epoch": 8.65, "grad_norm": 38.071407318115234, "learning_rate": 0.00021395118507114348, "loss": 1.7318, "step": 294500 }, { "epoch": 8.66, "grad_norm": 1.9748557806015015, "learning_rate": 0.0002138040930798121, "loss": 1.7327, "step": 295000 }, { "epoch": 8.68, "grad_norm": 1.8591829538345337, "learning_rate": 0.00021365700108848072, "loss": 1.6944, "step": 295500 }, { "epoch": 8.69, "grad_norm": 6.686271667480469, "learning_rate": 0.00021350990909714933, "loss": 1.7289, "step": 296000 }, { "epoch": 8.71, "grad_norm": 2.0325803756713867, "learning_rate": 0.00021336281710581794, "loss": 1.7074, "step": 296500 }, { "epoch": 8.72, "grad_norm": 2.5102651119232178, "learning_rate": 0.00021321572511448657, "loss": 1.7061, "step": 297000 }, { "epoch": 8.73, "grad_norm": 5.305497646331787, "learning_rate": 0.0002130686331231552, "loss": 1.7047, "step": 297500 }, { "epoch": 8.75, "grad_norm": 2.878256320953369, "learning_rate": 0.00021292154113182384, "loss": 1.7109, "step": 298000 }, { "epoch": 8.76, "grad_norm": 1.7807819843292236, "learning_rate": 0.00021277444914049245, "loss": 1.7136, "step": 298500 }, { "epoch": 8.78, "grad_norm": 1.8971943855285645, "learning_rate": 0.00021262735714916106, "loss": 1.7319, "step": 299000 }, { "epoch": 8.79, "grad_norm": 2.650453805923462, "learning_rate": 0.0002124802651578297, "loss": 1.7436, "step": 299500 }, { "epoch": 8.81, "grad_norm": 3.341278076171875, "learning_rate": 0.0002123331731664983, "loss": 1.6982, "step": 300000 }, { "epoch": 8.82, "grad_norm": 4.539509296417236, "learning_rate": 0.00021218608117516693, "loss": 1.7149, "step": 300500 }, { "epoch": 8.84, "grad_norm": 1.6525373458862305, "learning_rate": 0.00021203898918383554, "loss": 1.7154, "step": 301000 }, { "epoch": 8.85, "grad_norm": 2.1805810928344727, "learning_rate": 0.00021189189719250418, "loss": 1.7244, "step": 301500 }, { "epoch": 8.87, "grad_norm": 1.623687744140625, "learning_rate": 0.0002117448052011728, "loss": 1.7109, "step": 302000 }, { "epoch": 8.88, "grad_norm": 1.8576191663742065, "learning_rate": 0.00021159771320984142, "loss": 1.71, "step": 302500 }, { "epoch": 8.9, "grad_norm": 2.7003326416015625, "learning_rate": 0.00021145062121851005, "loss": 1.717, "step": 303000 }, { "epoch": 8.91, "grad_norm": 1.5555850267410278, "learning_rate": 0.00021130352922717866, "loss": 1.6771, "step": 303500 }, { "epoch": 8.93, "grad_norm": 1.8485440015792847, "learning_rate": 0.00021115643723584727, "loss": 1.6907, "step": 304000 }, { "epoch": 8.94, "grad_norm": 1.9917840957641602, "learning_rate": 0.0002110093452445159, "loss": 1.6929, "step": 304500 }, { "epoch": 8.95, "grad_norm": 2.54939866065979, "learning_rate": 0.0002108622532531845, "loss": 1.7211, "step": 305000 }, { "epoch": 8.97, "grad_norm": 4.3445234298706055, "learning_rate": 0.00021071516126185317, "loss": 1.7038, "step": 305500 }, { "epoch": 8.98, "grad_norm": 2.9025015830993652, "learning_rate": 0.00021056806927052178, "loss": 1.6987, "step": 306000 }, { "epoch": 9.0, "grad_norm": 3.951500177383423, "learning_rate": 0.0002104209772791904, "loss": 1.7063, "step": 306500 }, { "epoch": 9.01, "grad_norm": 2.2812907695770264, "learning_rate": 0.00021027388528785902, "loss": 1.6754, "step": 307000 }, { "epoch": 9.03, "grad_norm": 1.313977837562561, "learning_rate": 0.00021012679329652763, "loss": 1.6384, "step": 307500 }, { "epoch": 9.04, "grad_norm": 1.1051950454711914, "learning_rate": 0.00020997970130519624, "loss": 1.6372, "step": 308000 }, { "epoch": 9.06, "grad_norm": 2.6401305198669434, "learning_rate": 0.00020983260931386487, "loss": 1.6535, "step": 308500 }, { "epoch": 9.07, "grad_norm": 1.477562665939331, "learning_rate": 0.00020968551732253348, "loss": 1.6342, "step": 309000 }, { "epoch": 9.09, "grad_norm": 1.7968777418136597, "learning_rate": 0.00020953842533120214, "loss": 1.6163, "step": 309500 }, { "epoch": 9.1, "grad_norm": 1.8777967691421509, "learning_rate": 0.00020939133333987075, "loss": 1.6466, "step": 310000 }, { "epoch": 9.12, "grad_norm": 1.970590353012085, "learning_rate": 0.00020924424134853936, "loss": 1.6481, "step": 310500 }, { "epoch": 9.13, "grad_norm": 11.109952926635742, "learning_rate": 0.000209097149357208, "loss": 1.6634, "step": 311000 }, { "epoch": 9.15, "grad_norm": 1.5415191650390625, "learning_rate": 0.0002089500573658766, "loss": 1.637, "step": 311500 }, { "epoch": 9.16, "grad_norm": 2.2349135875701904, "learning_rate": 0.0002088029653745452, "loss": 1.6373, "step": 312000 }, { "epoch": 9.18, "grad_norm": 2.3896191120147705, "learning_rate": 0.00020865587338321384, "loss": 1.6318, "step": 312500 }, { "epoch": 9.19, "grad_norm": 1.5920429229736328, "learning_rate": 0.00020850878139188245, "loss": 1.6383, "step": 313000 }, { "epoch": 9.2, "grad_norm": 1.485474944114685, "learning_rate": 0.0002083616894005511, "loss": 1.6392, "step": 313500 }, { "epoch": 9.22, "grad_norm": 9.920788764953613, "learning_rate": 0.00020821459740921972, "loss": 1.6407, "step": 314000 }, { "epoch": 9.23, "grad_norm": 5.970518589019775, "learning_rate": 0.00020806750541788833, "loss": 1.6389, "step": 314500 }, { "epoch": 9.25, "grad_norm": 1.5371183156967163, "learning_rate": 0.00020792041342655696, "loss": 1.6485, "step": 315000 }, { "epoch": 9.26, "grad_norm": 2.242612600326538, "learning_rate": 0.00020777332143522557, "loss": 1.6344, "step": 315500 }, { "epoch": 9.28, "grad_norm": 3.137301445007324, "learning_rate": 0.00020762622944389418, "loss": 1.6269, "step": 316000 }, { "epoch": 9.29, "grad_norm": 3.391427993774414, "learning_rate": 0.0002074791374525628, "loss": 1.6575, "step": 316500 }, { "epoch": 9.31, "grad_norm": 7.92528772354126, "learning_rate": 0.00020733204546123142, "loss": 1.6333, "step": 317000 }, { "epoch": 9.32, "grad_norm": 1.0485262870788574, "learning_rate": 0.00020718495346990008, "loss": 1.6574, "step": 317500 }, { "epoch": 9.34, "grad_norm": 1.6511611938476562, "learning_rate": 0.0002070378614785687, "loss": 1.6474, "step": 318000 }, { "epoch": 9.35, "grad_norm": 1.3929743766784668, "learning_rate": 0.0002068907694872373, "loss": 1.6484, "step": 318500 }, { "epoch": 9.37, "grad_norm": 1.0993067026138306, "learning_rate": 0.00020674367749590593, "loss": 1.6312, "step": 319000 }, { "epoch": 9.38, "grad_norm": 2.8097622394561768, "learning_rate": 0.00020659658550457454, "loss": 1.6438, "step": 319500 }, { "epoch": 9.4, "grad_norm": 2.274348020553589, "learning_rate": 0.00020644949351324314, "loss": 1.6258, "step": 320000 }, { "epoch": 9.41, "grad_norm": 1.9707189798355103, "learning_rate": 0.00020630240152191178, "loss": 1.6619, "step": 320500 }, { "epoch": 9.42, "grad_norm": 2.7805778980255127, "learning_rate": 0.0002061553095305804, "loss": 1.6537, "step": 321000 }, { "epoch": 9.44, "grad_norm": 2.856705665588379, "learning_rate": 0.00020600821753924905, "loss": 1.6745, "step": 321500 }, { "epoch": 9.45, "grad_norm": 1.3102498054504395, "learning_rate": 0.00020586112554791766, "loss": 1.6662, "step": 322000 }, { "epoch": 9.47, "grad_norm": 4.1165690422058105, "learning_rate": 0.00020571403355658626, "loss": 1.6531, "step": 322500 }, { "epoch": 9.48, "grad_norm": 5.66050386428833, "learning_rate": 0.0002055669415652549, "loss": 1.669, "step": 323000 }, { "epoch": 9.5, "grad_norm": 1.3612234592437744, "learning_rate": 0.0002054198495739235, "loss": 1.642, "step": 323500 }, { "epoch": 9.51, "grad_norm": 2.5399467945098877, "learning_rate": 0.00020527275758259214, "loss": 1.677, "step": 324000 }, { "epoch": 9.53, "grad_norm": 5.590959548950195, "learning_rate": 0.00020512566559126075, "loss": 1.6673, "step": 324500 }, { "epoch": 9.54, "grad_norm": 1.6591867208480835, "learning_rate": 0.00020497857359992936, "loss": 1.6523, "step": 325000 }, { "epoch": 9.56, "grad_norm": 2.3825442790985107, "learning_rate": 0.00020483148160859802, "loss": 1.6472, "step": 325500 }, { "epoch": 9.57, "grad_norm": 1.509970784187317, "learning_rate": 0.00020468438961726663, "loss": 1.6095, "step": 326000 }, { "epoch": 9.59, "grad_norm": 1.0791665315628052, "learning_rate": 0.00020453729762593526, "loss": 1.6545, "step": 326500 }, { "epoch": 9.6, "grad_norm": 1.480979323387146, "learning_rate": 0.00020439020563460387, "loss": 1.6691, "step": 327000 }, { "epoch": 9.62, "grad_norm": 1.9298070669174194, "learning_rate": 0.00020424311364327248, "loss": 1.6561, "step": 327500 }, { "epoch": 9.63, "grad_norm": 2.1420977115631104, "learning_rate": 0.0002040960216519411, "loss": 1.6332, "step": 328000 }, { "epoch": 9.64, "grad_norm": 1.6322475671768188, "learning_rate": 0.00020394892966060972, "loss": 1.6576, "step": 328500 }, { "epoch": 9.66, "grad_norm": 1.7701492309570312, "learning_rate": 0.00020380183766927833, "loss": 1.6603, "step": 329000 }, { "epoch": 9.67, "grad_norm": 22.195663452148438, "learning_rate": 0.000203654745677947, "loss": 1.6605, "step": 329500 }, { "epoch": 9.69, "grad_norm": 1.735055685043335, "learning_rate": 0.0002035076536866156, "loss": 1.6728, "step": 330000 }, { "epoch": 9.7, "grad_norm": 1.3548341989517212, "learning_rate": 0.00020336056169528423, "loss": 1.6606, "step": 330500 }, { "epoch": 9.72, "grad_norm": 2.0401058197021484, "learning_rate": 0.00020321346970395284, "loss": 1.6687, "step": 331000 }, { "epoch": 9.73, "grad_norm": 16.856937408447266, "learning_rate": 0.00020306637771262144, "loss": 1.6369, "step": 331500 }, { "epoch": 9.75, "grad_norm": 3.8989596366882324, "learning_rate": 0.00020291928572129008, "loss": 1.6454, "step": 332000 }, { "epoch": 9.76, "grad_norm": 1.6723474264144897, "learning_rate": 0.0002027721937299587, "loss": 1.6723, "step": 332500 }, { "epoch": 9.78, "grad_norm": 1.7381031513214111, "learning_rate": 0.0002026251017386273, "loss": 1.6672, "step": 333000 }, { "epoch": 9.79, "grad_norm": 1.7013987302780151, "learning_rate": 0.00020247800974729596, "loss": 1.6666, "step": 333500 }, { "epoch": 9.81, "grad_norm": 1.6793595552444458, "learning_rate": 0.00020233091775596456, "loss": 1.6594, "step": 334000 }, { "epoch": 9.82, "grad_norm": 1.2351815700531006, "learning_rate": 0.0002021838257646332, "loss": 1.6475, "step": 334500 }, { "epoch": 9.84, "grad_norm": 3.5721399784088135, "learning_rate": 0.0002020367337733018, "loss": 1.6584, "step": 335000 }, { "epoch": 9.85, "grad_norm": 1.5881208181381226, "learning_rate": 0.0002018896417819704, "loss": 1.6584, "step": 335500 }, { "epoch": 9.87, "grad_norm": 2.2431886196136475, "learning_rate": 0.00020174254979063905, "loss": 1.6632, "step": 336000 }, { "epoch": 9.88, "grad_norm": 1.9759477376937866, "learning_rate": 0.00020159545779930766, "loss": 1.6098, "step": 336500 }, { "epoch": 9.89, "grad_norm": 2.4007914066314697, "learning_rate": 0.00020144836580797626, "loss": 1.6553, "step": 337000 }, { "epoch": 9.91, "grad_norm": 1.1646785736083984, "learning_rate": 0.00020130127381664492, "loss": 1.6917, "step": 337500 }, { "epoch": 9.92, "grad_norm": 2.7321279048919678, "learning_rate": 0.00020115418182531353, "loss": 1.6165, "step": 338000 }, { "epoch": 9.94, "grad_norm": 2.1226370334625244, "learning_rate": 0.00020100708983398217, "loss": 1.6583, "step": 338500 }, { "epoch": 9.95, "grad_norm": 16.655624389648438, "learning_rate": 0.00020085999784265077, "loss": 1.6594, "step": 339000 }, { "epoch": 9.97, "grad_norm": 6.142359256744385, "learning_rate": 0.00020071290585131938, "loss": 1.6343, "step": 339500 }, { "epoch": 9.98, "grad_norm": 2.1720430850982666, "learning_rate": 0.00020056581385998802, "loss": 1.6509, "step": 340000 }, { "epoch": 10.0, "grad_norm": 2.870986223220825, "learning_rate": 0.00020041872186865662, "loss": 1.6707, "step": 340500 }, { "epoch": 10.01, "grad_norm": 1.382546067237854, "learning_rate": 0.00020027162987732526, "loss": 1.5759, "step": 341000 }, { "epoch": 10.03, "grad_norm": 1.2603851556777954, "learning_rate": 0.0002001245378859939, "loss": 1.5937, "step": 341500 }, { "epoch": 10.04, "grad_norm": 2.455425262451172, "learning_rate": 0.0001999774458946625, "loss": 1.5859, "step": 342000 }, { "epoch": 10.06, "grad_norm": 1.6993873119354248, "learning_rate": 0.00019983035390333114, "loss": 1.6102, "step": 342500 }, { "epoch": 10.07, "grad_norm": 2.338402509689331, "learning_rate": 0.00019968326191199974, "loss": 1.6058, "step": 343000 }, { "epoch": 10.09, "grad_norm": 1.4525296688079834, "learning_rate": 0.00019953616992066838, "loss": 1.5925, "step": 343500 }, { "epoch": 10.1, "grad_norm": 1.7461038827896118, "learning_rate": 0.00019938907792933699, "loss": 1.5824, "step": 344000 }, { "epoch": 10.11, "grad_norm": 20.686992645263672, "learning_rate": 0.0001992419859380056, "loss": 1.6173, "step": 344500 }, { "epoch": 10.13, "grad_norm": 11.99207592010498, "learning_rate": 0.00019909489394667423, "loss": 1.5998, "step": 345000 }, { "epoch": 10.14, "grad_norm": 2.926992177963257, "learning_rate": 0.00019894780195534286, "loss": 1.5864, "step": 345500 }, { "epoch": 10.16, "grad_norm": 3.7215046882629395, "learning_rate": 0.0001988007099640115, "loss": 1.5856, "step": 346000 }, { "epoch": 10.17, "grad_norm": 1.7094500064849854, "learning_rate": 0.0001986536179726801, "loss": 1.5964, "step": 346500 }, { "epoch": 10.19, "grad_norm": 2.5829222202301025, "learning_rate": 0.0001985065259813487, "loss": 1.5943, "step": 347000 }, { "epoch": 10.2, "grad_norm": 1.6218924522399902, "learning_rate": 0.00019835943399001735, "loss": 1.5808, "step": 347500 }, { "epoch": 10.22, "grad_norm": 1.952506184577942, "learning_rate": 0.00019821234199868595, "loss": 1.5548, "step": 348000 }, { "epoch": 10.23, "grad_norm": 2.1660783290863037, "learning_rate": 0.00019806525000735456, "loss": 1.585, "step": 348500 }, { "epoch": 10.25, "grad_norm": 1.8295018672943115, "learning_rate": 0.0001979181580160232, "loss": 1.5975, "step": 349000 }, { "epoch": 10.26, "grad_norm": 2.3375236988067627, "learning_rate": 0.00019777106602469183, "loss": 1.617, "step": 349500 }, { "epoch": 10.28, "grad_norm": 1.323554515838623, "learning_rate": 0.00019762397403336047, "loss": 1.5915, "step": 350000 }, { "epoch": 10.29, "grad_norm": 1.259614109992981, "learning_rate": 0.00019747688204202907, "loss": 1.6126, "step": 350500 }, { "epoch": 10.31, "grad_norm": 27.243751525878906, "learning_rate": 0.00019732979005069768, "loss": 1.5886, "step": 351000 }, { "epoch": 10.32, "grad_norm": 1.4931347370147705, "learning_rate": 0.00019718269805936632, "loss": 1.6165, "step": 351500 }, { "epoch": 10.33, "grad_norm": 1.3573870658874512, "learning_rate": 0.00019703560606803492, "loss": 1.5967, "step": 352000 }, { "epoch": 10.35, "grad_norm": 2.393355369567871, "learning_rate": 0.00019688851407670353, "loss": 1.6033, "step": 352500 }, { "epoch": 10.36, "grad_norm": 17.975017547607422, "learning_rate": 0.00019674142208537217, "loss": 1.6176, "step": 353000 }, { "epoch": 10.38, "grad_norm": 2.195270538330078, "learning_rate": 0.0001965943300940408, "loss": 1.6022, "step": 353500 }, { "epoch": 10.39, "grad_norm": 1.5013370513916016, "learning_rate": 0.00019644723810270944, "loss": 1.5999, "step": 354000 }, { "epoch": 10.41, "grad_norm": 16.84026527404785, "learning_rate": 0.00019630014611137804, "loss": 1.6061, "step": 354500 }, { "epoch": 10.42, "grad_norm": 2.8341875076293945, "learning_rate": 0.00019615305412004665, "loss": 1.5742, "step": 355000 }, { "epoch": 10.44, "grad_norm": 1.6109588146209717, "learning_rate": 0.00019600596212871529, "loss": 1.5941, "step": 355500 }, { "epoch": 10.45, "grad_norm": 5.435859203338623, "learning_rate": 0.0001958588701373839, "loss": 1.5875, "step": 356000 }, { "epoch": 10.47, "grad_norm": 4.453019142150879, "learning_rate": 0.0001957117781460525, "loss": 1.5901, "step": 356500 }, { "epoch": 10.48, "grad_norm": 2.145413875579834, "learning_rate": 0.00019556468615472116, "loss": 1.5946, "step": 357000 }, { "epoch": 10.5, "grad_norm": 2.531158208847046, "learning_rate": 0.00019541759416338977, "loss": 1.5865, "step": 357500 }, { "epoch": 10.51, "grad_norm": 1.8745834827423096, "learning_rate": 0.0001952705021720584, "loss": 1.6033, "step": 358000 }, { "epoch": 10.53, "grad_norm": 1.3398380279541016, "learning_rate": 0.000195123410180727, "loss": 1.6205, "step": 358500 }, { "epoch": 10.54, "grad_norm": 6.710257053375244, "learning_rate": 0.00019497631818939562, "loss": 1.6171, "step": 359000 }, { "epoch": 10.56, "grad_norm": 1.5463893413543701, "learning_rate": 0.00019482922619806425, "loss": 1.5913, "step": 359500 }, { "epoch": 10.57, "grad_norm": 1.32403564453125, "learning_rate": 0.00019468213420673286, "loss": 1.6129, "step": 360000 }, { "epoch": 10.58, "grad_norm": 1.1276665925979614, "learning_rate": 0.00019453504221540147, "loss": 1.6063, "step": 360500 }, { "epoch": 10.6, "grad_norm": 3.2467503547668457, "learning_rate": 0.00019438795022407013, "loss": 1.6036, "step": 361000 }, { "epoch": 10.61, "grad_norm": 2.091618537902832, "learning_rate": 0.00019424085823273874, "loss": 1.5919, "step": 361500 }, { "epoch": 10.63, "grad_norm": 2.571202516555786, "learning_rate": 0.00019409376624140737, "loss": 1.6218, "step": 362000 }, { "epoch": 10.64, "grad_norm": 1.7798309326171875, "learning_rate": 0.00019394667425007598, "loss": 1.5919, "step": 362500 }, { "epoch": 10.66, "grad_norm": 1.96729576587677, "learning_rate": 0.0001937995822587446, "loss": 1.6116, "step": 363000 }, { "epoch": 10.67, "grad_norm": 3.229706048965454, "learning_rate": 0.00019365249026741322, "loss": 1.5993, "step": 363500 }, { "epoch": 10.69, "grad_norm": 6.615269184112549, "learning_rate": 0.00019350539827608183, "loss": 1.6034, "step": 364000 }, { "epoch": 10.7, "grad_norm": 6.847731113433838, "learning_rate": 0.00019335830628475047, "loss": 1.5785, "step": 364500 }, { "epoch": 10.72, "grad_norm": 1.3055931329727173, "learning_rate": 0.0001932112142934191, "loss": 1.6059, "step": 365000 }, { "epoch": 10.73, "grad_norm": 5.202936172485352, "learning_rate": 0.0001930641223020877, "loss": 1.6145, "step": 365500 }, { "epoch": 10.75, "grad_norm": 1.704176902770996, "learning_rate": 0.00019291703031075634, "loss": 1.6349, "step": 366000 }, { "epoch": 10.76, "grad_norm": 2.4655399322509766, "learning_rate": 0.00019276993831942495, "loss": 1.5845, "step": 366500 }, { "epoch": 10.78, "grad_norm": 6.769374370574951, "learning_rate": 0.00019262284632809358, "loss": 1.6054, "step": 367000 }, { "epoch": 10.79, "grad_norm": 3.4355087280273438, "learning_rate": 0.0001924757543367622, "loss": 1.6205, "step": 367500 }, { "epoch": 10.8, "grad_norm": 1.796799659729004, "learning_rate": 0.0001923286623454308, "loss": 1.6138, "step": 368000 }, { "epoch": 10.82, "grad_norm": 4.333281993865967, "learning_rate": 0.00019218157035409943, "loss": 1.6273, "step": 368500 }, { "epoch": 10.83, "grad_norm": 10.447500228881836, "learning_rate": 0.00019203447836276807, "loss": 1.6497, "step": 369000 }, { "epoch": 10.85, "grad_norm": 1.9053871631622314, "learning_rate": 0.0001918873863714367, "loss": 1.5815, "step": 369500 }, { "epoch": 10.86, "grad_norm": 6.775355339050293, "learning_rate": 0.0001917402943801053, "loss": 1.5986, "step": 370000 }, { "epoch": 10.88, "grad_norm": 19.00932502746582, "learning_rate": 0.00019159320238877392, "loss": 1.5916, "step": 370500 }, { "epoch": 10.89, "grad_norm": 2.1029250621795654, "learning_rate": 0.00019144611039744255, "loss": 1.6072, "step": 371000 }, { "epoch": 10.91, "grad_norm": 1.3043904304504395, "learning_rate": 0.00019129901840611116, "loss": 1.5885, "step": 371500 }, { "epoch": 10.92, "grad_norm": 1.8047374486923218, "learning_rate": 0.00019115192641477977, "loss": 1.6265, "step": 372000 }, { "epoch": 10.94, "grad_norm": 2.591125011444092, "learning_rate": 0.0001910048344234484, "loss": 1.6028, "step": 372500 }, { "epoch": 10.95, "grad_norm": 1.8858853578567505, "learning_rate": 0.00019085774243211704, "loss": 1.6212, "step": 373000 }, { "epoch": 10.97, "grad_norm": 1.4778261184692383, "learning_rate": 0.00019071065044078567, "loss": 1.6049, "step": 373500 }, { "epoch": 10.98, "grad_norm": 7.7040934562683105, "learning_rate": 0.00019056355844945428, "loss": 1.5926, "step": 374000 }, { "epoch": 11.0, "grad_norm": 1.5968307256698608, "learning_rate": 0.0001904164664581229, "loss": 1.6438, "step": 374500 }, { "epoch": 11.01, "grad_norm": 1.4240366220474243, "learning_rate": 0.00019026937446679152, "loss": 1.5451, "step": 375000 }, { "epoch": 11.02, "grad_norm": 2.0358214378356934, "learning_rate": 0.00019012228247546013, "loss": 1.5232, "step": 375500 }, { "epoch": 11.04, "grad_norm": 1.2968121767044067, "learning_rate": 0.00018997519048412874, "loss": 1.5417, "step": 376000 }, { "epoch": 11.05, "grad_norm": 3.3961257934570312, "learning_rate": 0.00018982809849279737, "loss": 1.5513, "step": 376500 }, { "epoch": 11.07, "grad_norm": 2.7855663299560547, "learning_rate": 0.000189681006501466, "loss": 1.5437, "step": 377000 }, { "epoch": 11.08, "grad_norm": 5.572605609893799, "learning_rate": 0.00018953391451013464, "loss": 1.5246, "step": 377500 }, { "epoch": 11.1, "grad_norm": 2.3603007793426514, "learning_rate": 0.00018938682251880325, "loss": 1.5246, "step": 378000 }, { "epoch": 11.11, "grad_norm": 19.196584701538086, "learning_rate": 0.00018923973052747186, "loss": 1.5223, "step": 378500 }, { "epoch": 11.13, "grad_norm": 2.4410743713378906, "learning_rate": 0.0001890926385361405, "loss": 1.5202, "step": 379000 }, { "epoch": 11.14, "grad_norm": 1.8335517644882202, "learning_rate": 0.0001889455465448091, "loss": 1.5625, "step": 379500 }, { "epoch": 11.16, "grad_norm": 2.2342159748077393, "learning_rate": 0.0001887984545534777, "loss": 1.5391, "step": 380000 }, { "epoch": 11.17, "grad_norm": 5.949936389923096, "learning_rate": 0.00018865136256214634, "loss": 1.569, "step": 380500 }, { "epoch": 11.19, "grad_norm": 6.047088623046875, "learning_rate": 0.00018850427057081498, "loss": 1.5366, "step": 381000 }, { "epoch": 11.2, "grad_norm": 16.05063819885254, "learning_rate": 0.0001883571785794836, "loss": 1.5343, "step": 381500 }, { "epoch": 11.22, "grad_norm": 1.327913761138916, "learning_rate": 0.00018821008658815222, "loss": 1.5641, "step": 382000 }, { "epoch": 11.23, "grad_norm": 4.132845878601074, "learning_rate": 0.00018806299459682083, "loss": 1.5722, "step": 382500 }, { "epoch": 11.25, "grad_norm": 1.9688639640808105, "learning_rate": 0.00018791590260548946, "loss": 1.5477, "step": 383000 }, { "epoch": 11.26, "grad_norm": 1.6657782793045044, "learning_rate": 0.00018776881061415807, "loss": 1.5398, "step": 383500 }, { "epoch": 11.27, "grad_norm": 2.9266579151153564, "learning_rate": 0.00018762171862282668, "loss": 1.556, "step": 384000 }, { "epoch": 11.29, "grad_norm": 5.755486011505127, "learning_rate": 0.0001874746266314953, "loss": 1.5415, "step": 384500 }, { "epoch": 11.3, "grad_norm": 2.1186070442199707, "learning_rate": 0.00018732753464016395, "loss": 1.5379, "step": 385000 }, { "epoch": 11.32, "grad_norm": 3.1836068630218506, "learning_rate": 0.00018718044264883258, "loss": 1.5621, "step": 385500 }, { "epoch": 11.33, "grad_norm": 4.884739875793457, "learning_rate": 0.0001870333506575012, "loss": 1.5791, "step": 386000 }, { "epoch": 11.35, "grad_norm": 2.4326329231262207, "learning_rate": 0.0001868862586661698, "loss": 1.5647, "step": 386500 }, { "epoch": 11.36, "grad_norm": 5.392803192138672, "learning_rate": 0.00018673916667483843, "loss": 1.5703, "step": 387000 }, { "epoch": 11.38, "grad_norm": 6.940735340118408, "learning_rate": 0.00018659207468350704, "loss": 1.5579, "step": 387500 }, { "epoch": 11.39, "grad_norm": 2.8021979331970215, "learning_rate": 0.00018644498269217567, "loss": 1.5469, "step": 388000 }, { "epoch": 11.41, "grad_norm": 1.8717076778411865, "learning_rate": 0.00018629789070084428, "loss": 1.5624, "step": 388500 }, { "epoch": 11.42, "grad_norm": 3.1023306846618652, "learning_rate": 0.00018615079870951291, "loss": 1.5751, "step": 389000 }, { "epoch": 11.44, "grad_norm": 1.5794864892959595, "learning_rate": 0.00018600370671818155, "loss": 1.5652, "step": 389500 }, { "epoch": 11.45, "grad_norm": 3.9795074462890625, "learning_rate": 0.00018585661472685016, "loss": 1.5663, "step": 390000 }, { "epoch": 11.47, "grad_norm": 1.4186744689941406, "learning_rate": 0.0001857095227355188, "loss": 1.5508, "step": 390500 }, { "epoch": 11.48, "grad_norm": 3.268982172012329, "learning_rate": 0.0001855624307441874, "loss": 1.5666, "step": 391000 }, { "epoch": 11.49, "grad_norm": 1.246903419494629, "learning_rate": 0.000185415338752856, "loss": 1.5797, "step": 391500 }, { "epoch": 11.51, "grad_norm": 1.4437174797058105, "learning_rate": 0.00018526824676152464, "loss": 1.562, "step": 392000 }, { "epoch": 11.52, "grad_norm": 1.8195850849151611, "learning_rate": 0.00018512115477019325, "loss": 1.5748, "step": 392500 }, { "epoch": 11.54, "grad_norm": 1.6077349185943604, "learning_rate": 0.0001849740627788619, "loss": 1.5628, "step": 393000 }, { "epoch": 11.55, "grad_norm": 1.614617943763733, "learning_rate": 0.00018482697078753052, "loss": 1.5992, "step": 393500 }, { "epoch": 11.57, "grad_norm": 2.860440731048584, "learning_rate": 0.00018467987879619913, "loss": 1.5635, "step": 394000 }, { "epoch": 11.58, "grad_norm": 4.223182201385498, "learning_rate": 0.00018453278680486776, "loss": 1.5678, "step": 394500 }, { "epoch": 11.6, "grad_norm": 2.143765687942505, "learning_rate": 0.00018438569481353637, "loss": 1.5714, "step": 395000 }, { "epoch": 11.61, "grad_norm": 14.07819652557373, "learning_rate": 0.00018423860282220498, "loss": 1.5662, "step": 395500 }, { "epoch": 11.63, "grad_norm": 1.6693955659866333, "learning_rate": 0.0001840915108308736, "loss": 1.5431, "step": 396000 }, { "epoch": 11.64, "grad_norm": 2.4783432483673096, "learning_rate": 0.00018394441883954222, "loss": 1.5439, "step": 396500 }, { "epoch": 11.66, "grad_norm": 2.1813952922821045, "learning_rate": 0.00018379732684821088, "loss": 1.5539, "step": 397000 }, { "epoch": 11.67, "grad_norm": 4.754596710205078, "learning_rate": 0.0001836502348568795, "loss": 1.5675, "step": 397500 }, { "epoch": 11.69, "grad_norm": 1.9912656545639038, "learning_rate": 0.0001835031428655481, "loss": 1.5909, "step": 398000 }, { "epoch": 11.7, "grad_norm": 1.5389857292175293, "learning_rate": 0.00018335605087421673, "loss": 1.5843, "step": 398500 }, { "epoch": 11.71, "grad_norm": 3.8105554580688477, "learning_rate": 0.00018320895888288534, "loss": 1.5804, "step": 399000 }, { "epoch": 11.73, "grad_norm": 2.6107397079467773, "learning_rate": 0.00018306186689155394, "loss": 1.5685, "step": 399500 }, { "epoch": 11.74, "grad_norm": 4.173858642578125, "learning_rate": 0.00018291477490022258, "loss": 1.5823, "step": 400000 }, { "epoch": 11.76, "grad_norm": 1.6713734865188599, "learning_rate": 0.0001827676829088912, "loss": 1.5672, "step": 400500 }, { "epoch": 11.77, "grad_norm": 4.650341033935547, "learning_rate": 0.00018262059091755985, "loss": 1.5297, "step": 401000 }, { "epoch": 11.79, "grad_norm": 1.5853548049926758, "learning_rate": 0.00018247349892622846, "loss": 1.5814, "step": 401500 }, { "epoch": 11.8, "grad_norm": 5.71536111831665, "learning_rate": 0.00018232640693489706, "loss": 1.5444, "step": 402000 }, { "epoch": 11.82, "grad_norm": 2.5603206157684326, "learning_rate": 0.0001821793149435657, "loss": 1.593, "step": 402500 }, { "epoch": 11.83, "grad_norm": 2.556755304336548, "learning_rate": 0.0001820322229522343, "loss": 1.5717, "step": 403000 }, { "epoch": 11.85, "grad_norm": 2.1441986560821533, "learning_rate": 0.0001818851309609029, "loss": 1.5766, "step": 403500 }, { "epoch": 11.86, "grad_norm": 1.762389898300171, "learning_rate": 0.00018173803896957155, "loss": 1.5857, "step": 404000 }, { "epoch": 11.88, "grad_norm": 1.7397152185440063, "learning_rate": 0.00018159094697824016, "loss": 1.5767, "step": 404500 }, { "epoch": 11.89, "grad_norm": 2.2762222290039062, "learning_rate": 0.00018144385498690882, "loss": 1.5735, "step": 405000 }, { "epoch": 11.91, "grad_norm": 2.9703893661499023, "learning_rate": 0.00018129676299557742, "loss": 1.5618, "step": 405500 }, { "epoch": 11.92, "grad_norm": 3.4913887977600098, "learning_rate": 0.00018114967100424603, "loss": 1.589, "step": 406000 }, { "epoch": 11.93, "grad_norm": 1.7856007814407349, "learning_rate": 0.00018100257901291467, "loss": 1.5924, "step": 406500 }, { "epoch": 11.95, "grad_norm": 5.869511604309082, "learning_rate": 0.00018085548702158327, "loss": 1.5516, "step": 407000 }, { "epoch": 11.96, "grad_norm": 1.4495360851287842, "learning_rate": 0.00018070839503025188, "loss": 1.5721, "step": 407500 }, { "epoch": 11.98, "grad_norm": 5.264606475830078, "learning_rate": 0.00018056130303892052, "loss": 1.5684, "step": 408000 }, { "epoch": 11.99, "grad_norm": 1.1457246541976929, "learning_rate": 0.00018041421104758912, "loss": 1.5747, "step": 408500 }, { "epoch": 12.01, "grad_norm": 2.9236228466033936, "learning_rate": 0.00018026711905625779, "loss": 1.5313, "step": 409000 }, { "epoch": 12.02, "grad_norm": 11.91308307647705, "learning_rate": 0.0001801200270649264, "loss": 1.4883, "step": 409500 }, { "epoch": 12.04, "grad_norm": 1.8393372297286987, "learning_rate": 0.000179972935073595, "loss": 1.4828, "step": 410000 }, { "epoch": 12.05, "grad_norm": 1.9591150283813477, "learning_rate": 0.00017982584308226364, "loss": 1.4832, "step": 410500 }, { "epoch": 12.07, "grad_norm": 1.5424935817718506, "learning_rate": 0.00017967875109093224, "loss": 1.4863, "step": 411000 }, { "epoch": 12.08, "grad_norm": 4.338646411895752, "learning_rate": 0.00017953165909960088, "loss": 1.5376, "step": 411500 }, { "epoch": 12.1, "grad_norm": 4.14846134185791, "learning_rate": 0.00017938456710826949, "loss": 1.506, "step": 412000 }, { "epoch": 12.11, "grad_norm": 6.690493106842041, "learning_rate": 0.00017923747511693812, "loss": 1.5015, "step": 412500 }, { "epoch": 12.13, "grad_norm": 4.801180362701416, "learning_rate": 0.00017909038312560676, "loss": 1.53, "step": 413000 }, { "epoch": 12.14, "grad_norm": 1.8997361660003662, "learning_rate": 0.00017894329113427536, "loss": 1.5191, "step": 413500 }, { "epoch": 12.16, "grad_norm": 19.748411178588867, "learning_rate": 0.000178796199142944, "loss": 1.5213, "step": 414000 }, { "epoch": 12.17, "grad_norm": 3.8095591068267822, "learning_rate": 0.0001786491071516126, "loss": 1.496, "step": 414500 }, { "epoch": 12.18, "grad_norm": 3.752877950668335, "learning_rate": 0.0001785020151602812, "loss": 1.5014, "step": 415000 }, { "epoch": 12.2, "grad_norm": 2.396458625793457, "learning_rate": 0.00017835492316894985, "loss": 1.4683, "step": 415500 }, { "epoch": 12.21, "grad_norm": 2.928061008453369, "learning_rate": 0.00017820783117761846, "loss": 1.5093, "step": 416000 }, { "epoch": 12.23, "grad_norm": 6.80102014541626, "learning_rate": 0.00017806073918628712, "loss": 1.5085, "step": 416500 }, { "epoch": 12.24, "grad_norm": 16.06849479675293, "learning_rate": 0.00017791364719495572, "loss": 1.5023, "step": 417000 }, { "epoch": 12.26, "grad_norm": 9.823871612548828, "learning_rate": 0.00017776655520362433, "loss": 1.5677, "step": 417500 }, { "epoch": 12.27, "grad_norm": 1.8364531993865967, "learning_rate": 0.00017761946321229297, "loss": 1.5146, "step": 418000 }, { "epoch": 12.29, "grad_norm": 2.6699376106262207, "learning_rate": 0.00017747237122096157, "loss": 1.5106, "step": 418500 }, { "epoch": 12.3, "grad_norm": 7.037272930145264, "learning_rate": 0.00017732527922963018, "loss": 1.5276, "step": 419000 }, { "epoch": 12.32, "grad_norm": 1.9254209995269775, "learning_rate": 0.00017717818723829882, "loss": 1.4907, "step": 419500 }, { "epoch": 12.33, "grad_norm": 5.405977249145508, "learning_rate": 0.00017703109524696742, "loss": 1.5039, "step": 420000 }, { "epoch": 12.35, "grad_norm": 6.036567211151123, "learning_rate": 0.00017688400325563609, "loss": 1.5258, "step": 420500 }, { "epoch": 12.36, "grad_norm": 3.601088523864746, "learning_rate": 0.0001767369112643047, "loss": 1.5316, "step": 421000 }, { "epoch": 12.38, "grad_norm": 3.247906446456909, "learning_rate": 0.0001765898192729733, "loss": 1.5006, "step": 421500 }, { "epoch": 12.39, "grad_norm": 2.138275384902954, "learning_rate": 0.00017644272728164194, "loss": 1.5125, "step": 422000 }, { "epoch": 12.4, "grad_norm": 2.024502992630005, "learning_rate": 0.00017629563529031054, "loss": 1.5323, "step": 422500 }, { "epoch": 12.42, "grad_norm": 21.49098777770996, "learning_rate": 0.00017614854329897915, "loss": 1.5607, "step": 423000 }, { "epoch": 12.43, "grad_norm": 3.5568199157714844, "learning_rate": 0.00017600145130764779, "loss": 1.4843, "step": 423500 }, { "epoch": 12.45, "grad_norm": 1.6709766387939453, "learning_rate": 0.0001758543593163164, "loss": 1.5621, "step": 424000 }, { "epoch": 12.46, "grad_norm": 2.4930756092071533, "learning_rate": 0.00017570726732498505, "loss": 1.5202, "step": 424500 }, { "epoch": 12.48, "grad_norm": 1.8919436931610107, "learning_rate": 0.00017556017533365366, "loss": 1.4859, "step": 425000 }, { "epoch": 12.49, "grad_norm": 1.2928180694580078, "learning_rate": 0.00017541308334232227, "loss": 1.5423, "step": 425500 }, { "epoch": 12.51, "grad_norm": 3.5068464279174805, "learning_rate": 0.0001752659913509909, "loss": 1.5326, "step": 426000 }, { "epoch": 12.52, "grad_norm": 2.7532618045806885, "learning_rate": 0.0001751188993596595, "loss": 1.5276, "step": 426500 }, { "epoch": 12.54, "grad_norm": 1.5163066387176514, "learning_rate": 0.00017497180736832812, "loss": 1.5306, "step": 427000 }, { "epoch": 12.55, "grad_norm": 21.672080993652344, "learning_rate": 0.00017482471537699675, "loss": 1.5018, "step": 427500 }, { "epoch": 12.57, "grad_norm": 4.169219970703125, "learning_rate": 0.00017467762338566536, "loss": 1.5314, "step": 428000 }, { "epoch": 12.58, "grad_norm": 1.7433300018310547, "learning_rate": 0.00017453053139433402, "loss": 1.5408, "step": 428500 }, { "epoch": 12.6, "grad_norm": 1.6866792440414429, "learning_rate": 0.00017438343940300263, "loss": 1.5405, "step": 429000 }, { "epoch": 12.61, "grad_norm": 12.107425689697266, "learning_rate": 0.00017423634741167124, "loss": 1.5375, "step": 429500 }, { "epoch": 12.62, "grad_norm": 2.0178098678588867, "learning_rate": 0.00017408925542033987, "loss": 1.5137, "step": 430000 }, { "epoch": 12.64, "grad_norm": 3.7658607959747314, "learning_rate": 0.00017394216342900848, "loss": 1.5138, "step": 430500 }, { "epoch": 12.65, "grad_norm": 2.0998480319976807, "learning_rate": 0.00017379507143767712, "loss": 1.5005, "step": 431000 }, { "epoch": 12.67, "grad_norm": 1.7215161323547363, "learning_rate": 0.00017364797944634572, "loss": 1.5083, "step": 431500 }, { "epoch": 12.68, "grad_norm": 3.4021968841552734, "learning_rate": 0.00017350088745501433, "loss": 1.5103, "step": 432000 }, { "epoch": 12.7, "grad_norm": 5.235382080078125, "learning_rate": 0.000173353795463683, "loss": 1.5491, "step": 432500 }, { "epoch": 12.71, "grad_norm": 5.455018997192383, "learning_rate": 0.0001732067034723516, "loss": 1.5486, "step": 433000 }, { "epoch": 12.73, "grad_norm": 2.4046807289123535, "learning_rate": 0.00017305961148102024, "loss": 1.5241, "step": 433500 }, { "epoch": 12.74, "grad_norm": 2.2005653381347656, "learning_rate": 0.00017291251948968884, "loss": 1.5392, "step": 434000 }, { "epoch": 12.76, "grad_norm": 1.7510486841201782, "learning_rate": 0.00017276542749835745, "loss": 1.5076, "step": 434500 }, { "epoch": 12.77, "grad_norm": 2.4069275856018066, "learning_rate": 0.00017261833550702608, "loss": 1.5189, "step": 435000 }, { "epoch": 12.79, "grad_norm": 3.2867445945739746, "learning_rate": 0.0001724712435156947, "loss": 1.5082, "step": 435500 }, { "epoch": 12.8, "grad_norm": 1.321496605873108, "learning_rate": 0.0001723241515243633, "loss": 1.5364, "step": 436000 }, { "epoch": 12.82, "grad_norm": 2.841768980026245, "learning_rate": 0.00017217705953303196, "loss": 1.5062, "step": 436500 }, { "epoch": 12.83, "grad_norm": 2.678611993789673, "learning_rate": 0.00017202996754170057, "loss": 1.5048, "step": 437000 }, { "epoch": 12.85, "grad_norm": 3.197082996368408, "learning_rate": 0.0001718828755503692, "loss": 1.5553, "step": 437500 }, { "epoch": 12.86, "grad_norm": 2.398364543914795, "learning_rate": 0.0001717357835590378, "loss": 1.4866, "step": 438000 }, { "epoch": 12.87, "grad_norm": 2.2497568130493164, "learning_rate": 0.00017158869156770642, "loss": 1.504, "step": 438500 }, { "epoch": 12.89, "grad_norm": 3.9616646766662598, "learning_rate": 0.00017144159957637505, "loss": 1.5049, "step": 439000 }, { "epoch": 12.9, "grad_norm": 1.5139656066894531, "learning_rate": 0.00017129450758504366, "loss": 1.549, "step": 439500 }, { "epoch": 12.92, "grad_norm": 2.305665969848633, "learning_rate": 0.00017114741559371227, "loss": 1.5255, "step": 440000 }, { "epoch": 12.93, "grad_norm": 2.3823301792144775, "learning_rate": 0.00017100032360238093, "loss": 1.4995, "step": 440500 }, { "epoch": 12.95, "grad_norm": 1.6001240015029907, "learning_rate": 0.00017085323161104954, "loss": 1.5494, "step": 441000 }, { "epoch": 12.96, "grad_norm": 2.843163013458252, "learning_rate": 0.00017070613961971817, "loss": 1.5394, "step": 441500 }, { "epoch": 12.98, "grad_norm": 2.1387860774993896, "learning_rate": 0.00017055904762838678, "loss": 1.5456, "step": 442000 }, { "epoch": 12.99, "grad_norm": 4.817657470703125, "learning_rate": 0.0001704119556370554, "loss": 1.5453, "step": 442500 }, { "epoch": 13.01, "grad_norm": 1.376753807067871, "learning_rate": 0.00017026486364572402, "loss": 1.5093, "step": 443000 }, { "epoch": 13.02, "grad_norm": 1.1947274208068848, "learning_rate": 0.00017011777165439263, "loss": 1.457, "step": 443500 }, { "epoch": 13.04, "grad_norm": 3.931847333908081, "learning_rate": 0.00016997067966306124, "loss": 1.4286, "step": 444000 }, { "epoch": 13.05, "grad_norm": 14.636221885681152, "learning_rate": 0.0001698235876717299, "loss": 1.4712, "step": 444500 }, { "epoch": 13.07, "grad_norm": 3.1448845863342285, "learning_rate": 0.0001696764956803985, "loss": 1.4628, "step": 445000 }, { "epoch": 13.08, "grad_norm": 0.9527159929275513, "learning_rate": 0.00016952940368906714, "loss": 1.4179, "step": 445500 }, { "epoch": 13.09, "grad_norm": 2.0560357570648193, "learning_rate": 0.00016938231169773575, "loss": 1.4716, "step": 446000 }, { "epoch": 13.11, "grad_norm": 2.8239283561706543, "learning_rate": 0.00016923521970640436, "loss": 1.477, "step": 446500 }, { "epoch": 13.12, "grad_norm": 3.246338367462158, "learning_rate": 0.000169088127715073, "loss": 1.4739, "step": 447000 }, { "epoch": 13.14, "grad_norm": 4.943697929382324, "learning_rate": 0.0001689410357237416, "loss": 1.4772, "step": 447500 }, { "epoch": 13.15, "grad_norm": 4.003148555755615, "learning_rate": 0.0001687939437324102, "loss": 1.4723, "step": 448000 }, { "epoch": 13.17, "grad_norm": 1.5625710487365723, "learning_rate": 0.00016864685174107887, "loss": 1.4424, "step": 448500 }, { "epoch": 13.18, "grad_norm": 2.3796448707580566, "learning_rate": 0.00016849975974974748, "loss": 1.512, "step": 449000 }, { "epoch": 13.2, "grad_norm": 2.314465284347534, "learning_rate": 0.0001683526677584161, "loss": 1.5012, "step": 449500 }, { "epoch": 13.21, "grad_norm": 1.8954664468765259, "learning_rate": 0.00016820557576708472, "loss": 1.4634, "step": 450000 }, { "epoch": 13.23, "grad_norm": 1.6504507064819336, "learning_rate": 0.00016805848377575333, "loss": 1.496, "step": 450500 }, { "epoch": 13.24, "grad_norm": 1.8685587644577026, "learning_rate": 0.00016791139178442196, "loss": 1.5054, "step": 451000 }, { "epoch": 13.26, "grad_norm": 3.0073273181915283, "learning_rate": 0.00016776429979309057, "loss": 1.4714, "step": 451500 }, { "epoch": 13.27, "grad_norm": 1.5322887897491455, "learning_rate": 0.0001676172078017592, "loss": 1.4683, "step": 452000 }, { "epoch": 13.29, "grad_norm": 10.244864463806152, "learning_rate": 0.00016747011581042784, "loss": 1.4834, "step": 452500 }, { "epoch": 13.3, "grad_norm": 2.7981672286987305, "learning_rate": 0.00016732302381909645, "loss": 1.4919, "step": 453000 }, { "epoch": 13.31, "grad_norm": 1.3389350175857544, "learning_rate": 0.00016717593182776508, "loss": 1.4589, "step": 453500 }, { "epoch": 13.33, "grad_norm": 2.1007466316223145, "learning_rate": 0.0001670288398364337, "loss": 1.4844, "step": 454000 }, { "epoch": 13.34, "grad_norm": 1.9225057363510132, "learning_rate": 0.00016688174784510232, "loss": 1.479, "step": 454500 }, { "epoch": 13.36, "grad_norm": 6.121819972991943, "learning_rate": 0.00016673465585377093, "loss": 1.4713, "step": 455000 }, { "epoch": 13.37, "grad_norm": 2.738173246383667, "learning_rate": 0.00016658756386243954, "loss": 1.5046, "step": 455500 }, { "epoch": 13.39, "grad_norm": 1.3760331869125366, "learning_rate": 0.00016644047187110817, "loss": 1.4919, "step": 456000 }, { "epoch": 13.4, "grad_norm": 9.234387397766113, "learning_rate": 0.0001662933798797768, "loss": 1.4531, "step": 456500 }, { "epoch": 13.42, "grad_norm": 1.6355879306793213, "learning_rate": 0.00016614628788844544, "loss": 1.442, "step": 457000 }, { "epoch": 13.43, "grad_norm": 5.203847885131836, "learning_rate": 0.00016599919589711405, "loss": 1.4672, "step": 457500 }, { "epoch": 13.45, "grad_norm": 5.590662956237793, "learning_rate": 0.00016585210390578266, "loss": 1.459, "step": 458000 }, { "epoch": 13.46, "grad_norm": 1.945825457572937, "learning_rate": 0.0001657050119144513, "loss": 1.4632, "step": 458500 }, { "epoch": 13.48, "grad_norm": 5.772292137145996, "learning_rate": 0.0001655579199231199, "loss": 1.4581, "step": 459000 }, { "epoch": 13.49, "grad_norm": 1.2538419961929321, "learning_rate": 0.0001654108279317885, "loss": 1.4735, "step": 459500 }, { "epoch": 13.51, "grad_norm": 5.774242401123047, "learning_rate": 0.00016526373594045714, "loss": 1.4833, "step": 460000 }, { "epoch": 13.52, "grad_norm": 1.8276828527450562, "learning_rate": 0.00016511664394912578, "loss": 1.5017, "step": 460500 }, { "epoch": 13.54, "grad_norm": 2.792278528213501, "learning_rate": 0.0001649695519577944, "loss": 1.4677, "step": 461000 }, { "epoch": 13.55, "grad_norm": 2.325228691101074, "learning_rate": 0.00016482245996646302, "loss": 1.482, "step": 461500 }, { "epoch": 13.56, "grad_norm": 2.3256492614746094, "learning_rate": 0.00016467536797513163, "loss": 1.4808, "step": 462000 }, { "epoch": 13.58, "grad_norm": 1.2064257860183716, "learning_rate": 0.00016452827598380026, "loss": 1.4695, "step": 462500 }, { "epoch": 13.59, "grad_norm": 3.7189693450927734, "learning_rate": 0.00016438118399246887, "loss": 1.4593, "step": 463000 }, { "epoch": 13.61, "grad_norm": 1.6200836896896362, "learning_rate": 0.00016423409200113748, "loss": 1.5176, "step": 463500 }, { "epoch": 13.62, "grad_norm": 3.3271920680999756, "learning_rate": 0.0001640870000098061, "loss": 1.4941, "step": 464000 }, { "epoch": 13.64, "grad_norm": 1.7106289863586426, "learning_rate": 0.00016393990801847474, "loss": 1.4414, "step": 464500 }, { "epoch": 13.65, "grad_norm": 2.1748921871185303, "learning_rate": 0.00016379281602714338, "loss": 1.4887, "step": 465000 }, { "epoch": 13.67, "grad_norm": 1.4012078046798706, "learning_rate": 0.000163645724035812, "loss": 1.4773, "step": 465500 }, { "epoch": 13.68, "grad_norm": 1.6043895483016968, "learning_rate": 0.0001634986320444806, "loss": 1.5182, "step": 466000 }, { "epoch": 13.7, "grad_norm": 1.53175950050354, "learning_rate": 0.00016335154005314923, "loss": 1.4939, "step": 466500 }, { "epoch": 13.71, "grad_norm": 7.83038854598999, "learning_rate": 0.00016320444806181784, "loss": 1.4997, "step": 467000 }, { "epoch": 13.73, "grad_norm": 20.90247344970703, "learning_rate": 0.00016305735607048644, "loss": 1.4717, "step": 467500 }, { "epoch": 13.74, "grad_norm": 1.7639943361282349, "learning_rate": 0.00016291026407915508, "loss": 1.4682, "step": 468000 }, { "epoch": 13.76, "grad_norm": 1.8345731496810913, "learning_rate": 0.00016276317208782371, "loss": 1.4941, "step": 468500 }, { "epoch": 13.77, "grad_norm": 13.04284381866455, "learning_rate": 0.00016261608009649235, "loss": 1.508, "step": 469000 }, { "epoch": 13.78, "grad_norm": 3.476304531097412, "learning_rate": 0.00016246898810516096, "loss": 1.4849, "step": 469500 }, { "epoch": 13.8, "grad_norm": 76.50082397460938, "learning_rate": 0.00016232189611382956, "loss": 1.5388, "step": 470000 }, { "epoch": 13.81, "grad_norm": 2.03076171875, "learning_rate": 0.0001621748041224982, "loss": 1.4964, "step": 470500 }, { "epoch": 13.83, "grad_norm": 5.736513137817383, "learning_rate": 0.0001620277121311668, "loss": 1.5211, "step": 471000 }, { "epoch": 13.84, "grad_norm": 1.7192264795303345, "learning_rate": 0.0001618806201398354, "loss": 1.4806, "step": 471500 }, { "epoch": 13.86, "grad_norm": 1.4231237173080444, "learning_rate": 0.00016173352814850408, "loss": 1.465, "step": 472000 }, { "epoch": 13.87, "grad_norm": 1.2601704597473145, "learning_rate": 0.00016158643615717268, "loss": 1.5081, "step": 472500 }, { "epoch": 13.89, "grad_norm": 1.2420154809951782, "learning_rate": 0.00016143934416584132, "loss": 1.5014, "step": 473000 }, { "epoch": 13.9, "grad_norm": 2.446563720703125, "learning_rate": 0.00016129225217450993, "loss": 1.4975, "step": 473500 }, { "epoch": 13.92, "grad_norm": 2.2312917709350586, "learning_rate": 0.00016114516018317853, "loss": 1.4827, "step": 474000 }, { "epoch": 13.93, "grad_norm": 2.2286717891693115, "learning_rate": 0.00016099806819184717, "loss": 1.4932, "step": 474500 }, { "epoch": 13.95, "grad_norm": 3.0326056480407715, "learning_rate": 0.00016085097620051577, "loss": 1.5015, "step": 475000 }, { "epoch": 13.96, "grad_norm": 4.350718975067139, "learning_rate": 0.0001607038842091844, "loss": 1.4917, "step": 475500 }, { "epoch": 13.98, "grad_norm": 1.2791180610656738, "learning_rate": 0.00016055679221785304, "loss": 1.5066, "step": 476000 }, { "epoch": 13.99, "grad_norm": 2.8099005222320557, "learning_rate": 0.00016040970022652165, "loss": 1.4746, "step": 476500 }, { "epoch": 14.0, "grad_norm": 2.087524175643921, "learning_rate": 0.0001602626082351903, "loss": 1.4278, "step": 477000 }, { "epoch": 14.02, "grad_norm": 5.598818778991699, "learning_rate": 0.0001601155162438589, "loss": 1.4266, "step": 477500 }, { "epoch": 14.03, "grad_norm": 1.573020100593567, "learning_rate": 0.00015996842425252753, "loss": 1.4198, "step": 478000 }, { "epoch": 14.05, "grad_norm": 2.8413734436035156, "learning_rate": 0.00015982133226119614, "loss": 1.421, "step": 478500 }, { "epoch": 14.06, "grad_norm": 1.3834174871444702, "learning_rate": 0.00015967424026986474, "loss": 1.4446, "step": 479000 }, { "epoch": 14.08, "grad_norm": 1.9328341484069824, "learning_rate": 0.00015952714827853338, "loss": 1.4332, "step": 479500 }, { "epoch": 14.09, "grad_norm": 4.331046104431152, "learning_rate": 0.000159380056287202, "loss": 1.4373, "step": 480000 }, { "epoch": 14.11, "grad_norm": 8.744680404663086, "learning_rate": 0.00015923296429587065, "loss": 1.4372, "step": 480500 }, { "epoch": 14.12, "grad_norm": 1.4437713623046875, "learning_rate": 0.00015908587230453926, "loss": 1.467, "step": 481000 }, { "epoch": 14.14, "grad_norm": 1.5478661060333252, "learning_rate": 0.00015893878031320786, "loss": 1.4209, "step": 481500 }, { "epoch": 14.15, "grad_norm": 12.385702133178711, "learning_rate": 0.0001587916883218765, "loss": 1.4529, "step": 482000 }, { "epoch": 14.17, "grad_norm": 1.585307002067566, "learning_rate": 0.0001586445963305451, "loss": 1.442, "step": 482500 }, { "epoch": 14.18, "grad_norm": 2.214268207550049, "learning_rate": 0.0001584975043392137, "loss": 1.4349, "step": 483000 }, { "epoch": 14.2, "grad_norm": 9.036513328552246, "learning_rate": 0.00015835041234788235, "loss": 1.4179, "step": 483500 }, { "epoch": 14.21, "grad_norm": 2.498725175857544, "learning_rate": 0.00015820332035655098, "loss": 1.4394, "step": 484000 }, { "epoch": 14.23, "grad_norm": 7.920095920562744, "learning_rate": 0.00015805622836521962, "loss": 1.4494, "step": 484500 }, { "epoch": 14.24, "grad_norm": 2.6044135093688965, "learning_rate": 0.00015790913637388822, "loss": 1.4159, "step": 485000 }, { "epoch": 14.25, "grad_norm": 1.906570315361023, "learning_rate": 0.00015776204438255683, "loss": 1.446, "step": 485500 }, { "epoch": 14.27, "grad_norm": 1.8452608585357666, "learning_rate": 0.00015761495239122547, "loss": 1.3961, "step": 486000 }, { "epoch": 14.28, "grad_norm": 1.6884217262268066, "learning_rate": 0.00015746786039989407, "loss": 1.4385, "step": 486500 }, { "epoch": 14.3, "grad_norm": 1.6386513710021973, "learning_rate": 0.00015732076840856268, "loss": 1.3842, "step": 487000 }, { "epoch": 14.31, "grad_norm": 5.906566619873047, "learning_rate": 0.00015717367641723132, "loss": 1.4241, "step": 487500 }, { "epoch": 14.33, "grad_norm": 2.964063882827759, "learning_rate": 0.00015702658442589995, "loss": 1.4454, "step": 488000 }, { "epoch": 14.34, "grad_norm": 1.53037428855896, "learning_rate": 0.00015687949243456859, "loss": 1.4067, "step": 488500 }, { "epoch": 14.36, "grad_norm": 1.5273714065551758, "learning_rate": 0.0001567324004432372, "loss": 1.4382, "step": 489000 }, { "epoch": 14.37, "grad_norm": 1.7097736597061157, "learning_rate": 0.0001565853084519058, "loss": 1.4167, "step": 489500 }, { "epoch": 14.39, "grad_norm": 2.170961380004883, "learning_rate": 0.00015643821646057444, "loss": 1.4541, "step": 490000 }, { "epoch": 14.4, "grad_norm": 5.699864387512207, "learning_rate": 0.00015629112446924304, "loss": 1.407, "step": 490500 }, { "epoch": 14.42, "grad_norm": 1.7430428266525269, "learning_rate": 0.00015614403247791165, "loss": 1.4418, "step": 491000 }, { "epoch": 14.43, "grad_norm": 2.179090976715088, "learning_rate": 0.00015599694048658029, "loss": 1.4331, "step": 491500 }, { "epoch": 14.45, "grad_norm": 1.534424901008606, "learning_rate": 0.00015584984849524892, "loss": 1.4536, "step": 492000 }, { "epoch": 14.46, "grad_norm": 1.464789867401123, "learning_rate": 0.00015570275650391755, "loss": 1.4459, "step": 492500 }, { "epoch": 14.47, "grad_norm": 1.9043675661087036, "learning_rate": 0.00015555566451258616, "loss": 1.4396, "step": 493000 }, { "epoch": 14.49, "grad_norm": 5.7054762840271, "learning_rate": 0.00015540857252125477, "loss": 1.4243, "step": 493500 }, { "epoch": 14.5, "grad_norm": 4.648599624633789, "learning_rate": 0.0001552614805299234, "loss": 1.4607, "step": 494000 }, { "epoch": 14.52, "grad_norm": 1.4880977869033813, "learning_rate": 0.000155114388538592, "loss": 1.4457, "step": 494500 }, { "epoch": 14.53, "grad_norm": 2.5349297523498535, "learning_rate": 0.00015496729654726062, "loss": 1.4675, "step": 495000 }, { "epoch": 14.55, "grad_norm": 3.5980210304260254, "learning_rate": 0.00015482020455592925, "loss": 1.4704, "step": 495500 }, { "epoch": 14.56, "grad_norm": 3.0215611457824707, "learning_rate": 0.0001546731125645979, "loss": 1.4202, "step": 496000 }, { "epoch": 14.58, "grad_norm": 1.6257708072662354, "learning_rate": 0.00015452602057326652, "loss": 1.4569, "step": 496500 }, { "epoch": 14.59, "grad_norm": 3.856419324874878, "learning_rate": 0.00015437892858193513, "loss": 1.4142, "step": 497000 }, { "epoch": 14.61, "grad_norm": 4.544717788696289, "learning_rate": 0.00015423183659060374, "loss": 1.4437, "step": 497500 }, { "epoch": 14.62, "grad_norm": 3.5249671936035156, "learning_rate": 0.00015408474459927237, "loss": 1.4432, "step": 498000 }, { "epoch": 14.64, "grad_norm": 1.8191946744918823, "learning_rate": 0.00015393765260794098, "loss": 1.435, "step": 498500 }, { "epoch": 14.65, "grad_norm": 2.6009254455566406, "learning_rate": 0.00015379056061660962, "loss": 1.411, "step": 499000 }, { "epoch": 14.67, "grad_norm": 2.475637674331665, "learning_rate": 0.00015364346862527822, "loss": 1.4568, "step": 499500 }, { "epoch": 14.68, "grad_norm": 3.965743064880371, "learning_rate": 0.00015349637663394686, "loss": 1.455, "step": 500000 }, { "epoch": 14.69, "grad_norm": 1.3109550476074219, "learning_rate": 0.0001533492846426155, "loss": 1.4547, "step": 500500 }, { "epoch": 14.71, "grad_norm": 2.0032811164855957, "learning_rate": 0.0001532021926512841, "loss": 1.4451, "step": 501000 }, { "epoch": 14.72, "grad_norm": 1.4996896982192993, "learning_rate": 0.00015305510065995274, "loss": 1.4497, "step": 501500 }, { "epoch": 14.74, "grad_norm": 2.1643595695495605, "learning_rate": 0.00015290800866862134, "loss": 1.4335, "step": 502000 }, { "epoch": 14.75, "grad_norm": 1.5510461330413818, "learning_rate": 0.00015276091667728995, "loss": 1.44, "step": 502500 }, { "epoch": 14.77, "grad_norm": 1.9841787815093994, "learning_rate": 0.00015261382468595858, "loss": 1.392, "step": 503000 }, { "epoch": 14.78, "grad_norm": 7.774050712585449, "learning_rate": 0.0001524667326946272, "loss": 1.4521, "step": 503500 }, { "epoch": 14.8, "grad_norm": 1.3330835103988647, "learning_rate": 0.00015231964070329585, "loss": 1.4562, "step": 504000 }, { "epoch": 14.81, "grad_norm": 2.6969919204711914, "learning_rate": 0.00015217254871196446, "loss": 1.4313, "step": 504500 }, { "epoch": 14.83, "grad_norm": 1.6068675518035889, "learning_rate": 0.00015202545672063307, "loss": 1.4515, "step": 505000 }, { "epoch": 14.84, "grad_norm": 3.470465660095215, "learning_rate": 0.0001518783647293017, "loss": 1.418, "step": 505500 }, { "epoch": 14.86, "grad_norm": 5.018684387207031, "learning_rate": 0.0001517312727379703, "loss": 1.4236, "step": 506000 }, { "epoch": 14.87, "grad_norm": 1.603996753692627, "learning_rate": 0.00015158418074663892, "loss": 1.4478, "step": 506500 }, { "epoch": 14.89, "grad_norm": 2.150404930114746, "learning_rate": 0.00015143708875530755, "loss": 1.4523, "step": 507000 }, { "epoch": 14.9, "grad_norm": 1.252964735031128, "learning_rate": 0.00015128999676397616, "loss": 1.4546, "step": 507500 }, { "epoch": 14.92, "grad_norm": 3.1075639724731445, "learning_rate": 0.00015114290477264482, "loss": 1.4547, "step": 508000 }, { "epoch": 14.93, "grad_norm": 1.4597883224487305, "learning_rate": 0.00015099581278131343, "loss": 1.4902, "step": 508500 }, { "epoch": 14.94, "grad_norm": 2.405596971511841, "learning_rate": 0.00015084872078998204, "loss": 1.4306, "step": 509000 }, { "epoch": 14.96, "grad_norm": 1.952222228050232, "learning_rate": 0.00015070162879865067, "loss": 1.4209, "step": 509500 }, { "epoch": 14.97, "grad_norm": 2.3458521366119385, "learning_rate": 0.00015055453680731928, "loss": 1.468, "step": 510000 }, { "epoch": 14.99, "grad_norm": 1.6442558765411377, "learning_rate": 0.0001504074448159879, "loss": 1.4541, "step": 510500 }, { "epoch": 15.0, "grad_norm": 2.1197829246520996, "learning_rate": 0.00015026035282465652, "loss": 1.4195, "step": 511000 }, { "epoch": 15.02, "grad_norm": 2.8808157444000244, "learning_rate": 0.00015011326083332513, "loss": 1.3719, "step": 511500 }, { "epoch": 15.03, "grad_norm": 1.939942717552185, "learning_rate": 0.00014996616884199377, "loss": 1.3627, "step": 512000 }, { "epoch": 15.05, "grad_norm": 4.798866271972656, "learning_rate": 0.00014981907685066237, "loss": 1.3839, "step": 512500 }, { "epoch": 15.06, "grad_norm": 2.553893566131592, "learning_rate": 0.000149671984859331, "loss": 1.4047, "step": 513000 }, { "epoch": 15.08, "grad_norm": 1.915277123451233, "learning_rate": 0.00014952489286799964, "loss": 1.3933, "step": 513500 }, { "epoch": 15.09, "grad_norm": 1.5252777338027954, "learning_rate": 0.00014937780087666825, "loss": 1.3644, "step": 514000 }, { "epoch": 15.11, "grad_norm": 1.6960341930389404, "learning_rate": 0.00014923070888533686, "loss": 1.3946, "step": 514500 }, { "epoch": 15.12, "grad_norm": 9.384785652160645, "learning_rate": 0.0001490836168940055, "loss": 1.3706, "step": 515000 }, { "epoch": 15.14, "grad_norm": 2.280630350112915, "learning_rate": 0.00014893652490267413, "loss": 1.3836, "step": 515500 }, { "epoch": 15.15, "grad_norm": 1.8966234922409058, "learning_rate": 0.00014878943291134273, "loss": 1.373, "step": 516000 }, { "epoch": 15.16, "grad_norm": 2.4201481342315674, "learning_rate": 0.00014864234092001134, "loss": 1.4014, "step": 516500 }, { "epoch": 15.18, "grad_norm": 2.8996338844299316, "learning_rate": 0.00014849524892867998, "loss": 1.3985, "step": 517000 }, { "epoch": 15.19, "grad_norm": 2.2007150650024414, "learning_rate": 0.0001483481569373486, "loss": 1.3823, "step": 517500 }, { "epoch": 15.21, "grad_norm": 2.7956199645996094, "learning_rate": 0.00014820106494601722, "loss": 1.3936, "step": 518000 }, { "epoch": 15.22, "grad_norm": 2.891369104385376, "learning_rate": 0.00014805397295468585, "loss": 1.3991, "step": 518500 }, { "epoch": 15.24, "grad_norm": 7.358971118927002, "learning_rate": 0.00014790688096335446, "loss": 1.4115, "step": 519000 }, { "epoch": 15.25, "grad_norm": 2.1062731742858887, "learning_rate": 0.0001477597889720231, "loss": 1.398, "step": 519500 }, { "epoch": 15.27, "grad_norm": 3.914013385772705, "learning_rate": 0.0001476126969806917, "loss": 1.4129, "step": 520000 }, { "epoch": 15.28, "grad_norm": 1.342411756515503, "learning_rate": 0.00014746560498936034, "loss": 1.4022, "step": 520500 }, { "epoch": 15.3, "grad_norm": 1.3836804628372192, "learning_rate": 0.00014731851299802897, "loss": 1.3849, "step": 521000 }, { "epoch": 15.31, "grad_norm": 4.170617580413818, "learning_rate": 0.00014717142100669758, "loss": 1.3855, "step": 521500 }, { "epoch": 15.33, "grad_norm": 6.060724258422852, "learning_rate": 0.0001470243290153662, "loss": 1.3645, "step": 522000 }, { "epoch": 15.34, "grad_norm": 5.009141445159912, "learning_rate": 0.00014687723702403482, "loss": 1.4106, "step": 522500 }, { "epoch": 15.36, "grad_norm": 4.241628646850586, "learning_rate": 0.00014673014503270346, "loss": 1.3983, "step": 523000 }, { "epoch": 15.37, "grad_norm": 1.6837831735610962, "learning_rate": 0.00014658305304137206, "loss": 1.4023, "step": 523500 }, { "epoch": 15.38, "grad_norm": 2.455502510070801, "learning_rate": 0.00014643596105004067, "loss": 1.3607, "step": 524000 }, { "epoch": 15.4, "grad_norm": 6.113306522369385, "learning_rate": 0.0001462888690587093, "loss": 1.4292, "step": 524500 }, { "epoch": 15.41, "grad_norm": 1.7169700860977173, "learning_rate": 0.00014614177706737794, "loss": 1.364, "step": 525000 }, { "epoch": 15.43, "grad_norm": 3.563976287841797, "learning_rate": 0.00014599468507604655, "loss": 1.4007, "step": 525500 }, { "epoch": 15.44, "grad_norm": 3.6780059337615967, "learning_rate": 0.00014584759308471516, "loss": 1.4072, "step": 526000 }, { "epoch": 15.46, "grad_norm": 1.8734827041625977, "learning_rate": 0.0001457005010933838, "loss": 1.4088, "step": 526500 }, { "epoch": 15.47, "grad_norm": 6.644404411315918, "learning_rate": 0.00014555340910205243, "loss": 1.3958, "step": 527000 }, { "epoch": 15.49, "grad_norm": 1.7657749652862549, "learning_rate": 0.00014540631711072103, "loss": 1.418, "step": 527500 }, { "epoch": 15.5, "grad_norm": 4.092434406280518, "learning_rate": 0.00014525922511938964, "loss": 1.3643, "step": 528000 }, { "epoch": 15.52, "grad_norm": 30.432186126708984, "learning_rate": 0.00014511213312805828, "loss": 1.3797, "step": 528500 }, { "epoch": 15.53, "grad_norm": 10.275367736816406, "learning_rate": 0.0001449650411367269, "loss": 1.3928, "step": 529000 }, { "epoch": 15.55, "grad_norm": 20.94750213623047, "learning_rate": 0.00014481794914539552, "loss": 1.4131, "step": 529500 }, { "epoch": 15.56, "grad_norm": 1.7463383674621582, "learning_rate": 0.00014467085715406413, "loss": 1.4173, "step": 530000 }, { "epoch": 15.58, "grad_norm": 2.380938768386841, "learning_rate": 0.00014452376516273276, "loss": 1.4092, "step": 530500 }, { "epoch": 15.59, "grad_norm": 1.8386043310165405, "learning_rate": 0.0001443766731714014, "loss": 1.3972, "step": 531000 }, { "epoch": 15.61, "grad_norm": 1.3829760551452637, "learning_rate": 0.00014422958118007, "loss": 1.391, "step": 531500 }, { "epoch": 15.62, "grad_norm": 2.171069383621216, "learning_rate": 0.0001440824891887386, "loss": 1.3993, "step": 532000 }, { "epoch": 15.63, "grad_norm": 1.716299057006836, "learning_rate": 0.00014393539719740724, "loss": 1.398, "step": 532500 }, { "epoch": 15.65, "grad_norm": 1.7643611431121826, "learning_rate": 0.00014378830520607588, "loss": 1.3901, "step": 533000 }, { "epoch": 15.66, "grad_norm": 1.68152916431427, "learning_rate": 0.0001436412132147445, "loss": 1.3873, "step": 533500 }, { "epoch": 15.68, "grad_norm": 2.581348419189453, "learning_rate": 0.0001434941212234131, "loss": 1.3713, "step": 534000 }, { "epoch": 15.69, "grad_norm": 3.0933191776275635, "learning_rate": 0.00014334702923208173, "loss": 1.4283, "step": 534500 }, { "epoch": 15.71, "grad_norm": 6.795374870300293, "learning_rate": 0.00014319993724075036, "loss": 1.4088, "step": 535000 }, { "epoch": 15.72, "grad_norm": 2.514035701751709, "learning_rate": 0.00014305284524941897, "loss": 1.4042, "step": 535500 }, { "epoch": 15.74, "grad_norm": 2.5651772022247314, "learning_rate": 0.00014290575325808758, "loss": 1.4236, "step": 536000 }, { "epoch": 15.75, "grad_norm": 1.6859495639801025, "learning_rate": 0.00014275866126675621, "loss": 1.4243, "step": 536500 }, { "epoch": 15.77, "grad_norm": 1.8449592590332031, "learning_rate": 0.00014261156927542485, "loss": 1.406, "step": 537000 }, { "epoch": 15.78, "grad_norm": 2.2886695861816406, "learning_rate": 0.00014246447728409346, "loss": 1.4095, "step": 537500 }, { "epoch": 15.8, "grad_norm": 2.669768810272217, "learning_rate": 0.00014231738529276206, "loss": 1.4172, "step": 538000 }, { "epoch": 15.81, "grad_norm": 5.082691192626953, "learning_rate": 0.0001421702933014307, "loss": 1.4307, "step": 538500 }, { "epoch": 15.83, "grad_norm": 2.9871368408203125, "learning_rate": 0.00014202320131009933, "loss": 1.3937, "step": 539000 }, { "epoch": 15.84, "grad_norm": 1.900804877281189, "learning_rate": 0.00014187610931876794, "loss": 1.4226, "step": 539500 }, { "epoch": 15.85, "grad_norm": 1.668407678604126, "learning_rate": 0.00014172901732743655, "loss": 1.3807, "step": 540000 }, { "epoch": 15.87, "grad_norm": 5.046024799346924, "learning_rate": 0.00014158192533610518, "loss": 1.4203, "step": 540500 }, { "epoch": 15.88, "grad_norm": 2.8824052810668945, "learning_rate": 0.00014143483334477382, "loss": 1.4188, "step": 541000 }, { "epoch": 15.9, "grad_norm": 2.688316583633423, "learning_rate": 0.00014128774135344243, "loss": 1.418, "step": 541500 }, { "epoch": 15.91, "grad_norm": 2.323672294616699, "learning_rate": 0.00014114064936211106, "loss": 1.3825, "step": 542000 }, { "epoch": 15.93, "grad_norm": 23.119873046875, "learning_rate": 0.00014099355737077967, "loss": 1.4148, "step": 542500 }, { "epoch": 15.94, "grad_norm": 3.257922649383545, "learning_rate": 0.0001408464653794483, "loss": 1.4327, "step": 543000 }, { "epoch": 15.96, "grad_norm": 1.8719940185546875, "learning_rate": 0.0001406993733881169, "loss": 1.3868, "step": 543500 }, { "epoch": 15.97, "grad_norm": 2.128316640853882, "learning_rate": 0.00014055228139678554, "loss": 1.4102, "step": 544000 }, { "epoch": 15.99, "grad_norm": 2.277371644973755, "learning_rate": 0.00014040518940545418, "loss": 1.4014, "step": 544500 }, { "epoch": 16.0, "grad_norm": 41.53245162963867, "learning_rate": 0.0001402580974141228, "loss": 1.4178, "step": 545000 }, { "epoch": 16.02, "grad_norm": 1.792492151260376, "learning_rate": 0.0001401110054227914, "loss": 1.3353, "step": 545500 }, { "epoch": 16.03, "grad_norm": 1.582220435142517, "learning_rate": 0.00013996391343146003, "loss": 1.3722, "step": 546000 }, { "epoch": 16.05, "grad_norm": 3.7628824710845947, "learning_rate": 0.00013981682144012866, "loss": 1.3593, "step": 546500 }, { "epoch": 16.06, "grad_norm": 2.833401918411255, "learning_rate": 0.00013966972944879727, "loss": 1.3627, "step": 547000 }, { "epoch": 16.07, "grad_norm": 2.1202030181884766, "learning_rate": 0.00013952263745746588, "loss": 1.3528, "step": 547500 }, { "epoch": 16.09, "grad_norm": 1.919555425643921, "learning_rate": 0.0001393755454661345, "loss": 1.3087, "step": 548000 }, { "epoch": 16.1, "grad_norm": 2.3834056854248047, "learning_rate": 0.00013922845347480315, "loss": 1.3303, "step": 548500 }, { "epoch": 16.12, "grad_norm": 2.0871472358703613, "learning_rate": 0.00013908136148347176, "loss": 1.3423, "step": 549000 }, { "epoch": 16.13, "grad_norm": 3.040555953979492, "learning_rate": 0.00013893426949214036, "loss": 1.3402, "step": 549500 }, { "epoch": 16.15, "grad_norm": 2.8366496562957764, "learning_rate": 0.000138787177500809, "loss": 1.3586, "step": 550000 }, { "epoch": 16.16, "grad_norm": 10.008976936340332, "learning_rate": 0.00013864008550947763, "loss": 1.3321, "step": 550500 }, { "epoch": 16.18, "grad_norm": 8.792502403259277, "learning_rate": 0.00013849299351814624, "loss": 1.3781, "step": 551000 }, { "epoch": 16.19, "grad_norm": 8.872962951660156, "learning_rate": 0.00013834590152681485, "loss": 1.3687, "step": 551500 }, { "epoch": 16.21, "grad_norm": 5.631560802459717, "learning_rate": 0.00013819880953548348, "loss": 1.3419, "step": 552000 }, { "epoch": 16.22, "grad_norm": 2.04437255859375, "learning_rate": 0.00013805171754415212, "loss": 1.3335, "step": 552500 }, { "epoch": 16.24, "grad_norm": 2.526149272918701, "learning_rate": 0.00013790462555282072, "loss": 1.3446, "step": 553000 }, { "epoch": 16.25, "grad_norm": 8.970195770263672, "learning_rate": 0.00013775753356148933, "loss": 1.3795, "step": 553500 }, { "epoch": 16.27, "grad_norm": 4.344628810882568, "learning_rate": 0.00013761044157015797, "loss": 1.3231, "step": 554000 }, { "epoch": 16.28, "grad_norm": 2.5846548080444336, "learning_rate": 0.0001374633495788266, "loss": 1.3571, "step": 554500 }, { "epoch": 16.3, "grad_norm": 20.92795181274414, "learning_rate": 0.0001373162575874952, "loss": 1.38, "step": 555000 }, { "epoch": 16.31, "grad_norm": 2.3088529109954834, "learning_rate": 0.00013716916559616382, "loss": 1.3473, "step": 555500 }, { "epoch": 16.32, "grad_norm": 13.217586517333984, "learning_rate": 0.00013702207360483245, "loss": 1.3558, "step": 556000 }, { "epoch": 16.34, "grad_norm": 8.642449378967285, "learning_rate": 0.00013687498161350109, "loss": 1.382, "step": 556500 }, { "epoch": 16.35, "grad_norm": 1.4824799299240112, "learning_rate": 0.0001367278896221697, "loss": 1.384, "step": 557000 }, { "epoch": 16.37, "grad_norm": 1.741585373878479, "learning_rate": 0.0001365807976308383, "loss": 1.3664, "step": 557500 }, { "epoch": 16.38, "grad_norm": 1.7038291692733765, "learning_rate": 0.00013643370563950694, "loss": 1.3634, "step": 558000 }, { "epoch": 16.4, "grad_norm": 17.11383628845215, "learning_rate": 0.00013628661364817557, "loss": 1.3604, "step": 558500 }, { "epoch": 16.41, "grad_norm": 19.561166763305664, "learning_rate": 0.00013613952165684418, "loss": 1.3731, "step": 559000 }, { "epoch": 16.43, "grad_norm": 2.2067980766296387, "learning_rate": 0.00013599242966551279, "loss": 1.3472, "step": 559500 }, { "epoch": 16.44, "grad_norm": 3.9638609886169434, "learning_rate": 0.00013584533767418142, "loss": 1.365, "step": 560000 }, { "epoch": 16.46, "grad_norm": 2.2947542667388916, "learning_rate": 0.00013569824568285006, "loss": 1.3618, "step": 560500 }, { "epoch": 16.47, "grad_norm": 2.3892598152160645, "learning_rate": 0.00013555115369151866, "loss": 1.3331, "step": 561000 }, { "epoch": 16.49, "grad_norm": 1.9236092567443848, "learning_rate": 0.00013540406170018727, "loss": 1.3508, "step": 561500 }, { "epoch": 16.5, "grad_norm": 5.180337429046631, "learning_rate": 0.0001352569697088559, "loss": 1.3612, "step": 562000 }, { "epoch": 16.52, "grad_norm": 8.786672592163086, "learning_rate": 0.00013510987771752454, "loss": 1.3776, "step": 562500 }, { "epoch": 16.53, "grad_norm": 4.111878871917725, "learning_rate": 0.00013496278572619315, "loss": 1.3661, "step": 563000 }, { "epoch": 16.54, "grad_norm": 4.706780433654785, "learning_rate": 0.00013481569373486178, "loss": 1.3731, "step": 563500 }, { "epoch": 16.56, "grad_norm": 1.3788596391677856, "learning_rate": 0.0001346686017435304, "loss": 1.3316, "step": 564000 }, { "epoch": 16.57, "grad_norm": 2.970449924468994, "learning_rate": 0.00013452150975219902, "loss": 1.3612, "step": 564500 }, { "epoch": 16.59, "grad_norm": 2.0503463745117188, "learning_rate": 0.00013437441776086763, "loss": 1.3635, "step": 565000 }, { "epoch": 16.6, "grad_norm": 10.668388366699219, "learning_rate": 0.00013422732576953627, "loss": 1.3764, "step": 565500 }, { "epoch": 16.62, "grad_norm": 8.57248592376709, "learning_rate": 0.00013408023377820487, "loss": 1.3521, "step": 566000 }, { "epoch": 16.63, "grad_norm": 1.6269396543502808, "learning_rate": 0.0001339331417868735, "loss": 1.3406, "step": 566500 }, { "epoch": 16.65, "grad_norm": 1.764863133430481, "learning_rate": 0.00013378604979554212, "loss": 1.3325, "step": 567000 }, { "epoch": 16.66, "grad_norm": 1.6542813777923584, "learning_rate": 0.00013363895780421075, "loss": 1.3601, "step": 567500 }, { "epoch": 16.68, "grad_norm": 1.9206827878952026, "learning_rate": 0.00013349186581287936, "loss": 1.3515, "step": 568000 }, { "epoch": 16.69, "grad_norm": 2.092914581298828, "learning_rate": 0.000133344773821548, "loss": 1.3617, "step": 568500 }, { "epoch": 16.71, "grad_norm": 6.570430278778076, "learning_rate": 0.0001331976818302166, "loss": 1.3728, "step": 569000 }, { "epoch": 16.72, "grad_norm": 2.719400644302368, "learning_rate": 0.00013305058983888524, "loss": 1.3406, "step": 569500 }, { "epoch": 16.74, "grad_norm": 9.820898056030273, "learning_rate": 0.00013290349784755384, "loss": 1.4036, "step": 570000 }, { "epoch": 16.75, "grad_norm": 2.3756299018859863, "learning_rate": 0.00013275640585622248, "loss": 1.3735, "step": 570500 }, { "epoch": 16.76, "grad_norm": 6.200007438659668, "learning_rate": 0.00013260931386489109, "loss": 1.3792, "step": 571000 }, { "epoch": 16.78, "grad_norm": 7.884439468383789, "learning_rate": 0.00013246222187355972, "loss": 1.3666, "step": 571500 }, { "epoch": 16.79, "grad_norm": 2.011915683746338, "learning_rate": 0.00013231512988222833, "loss": 1.3541, "step": 572000 }, { "epoch": 16.81, "grad_norm": 6.167238712310791, "learning_rate": 0.00013216803789089696, "loss": 1.3781, "step": 572500 }, { "epoch": 16.82, "grad_norm": 38.23750686645508, "learning_rate": 0.00013202094589956557, "loss": 1.3546, "step": 573000 }, { "epoch": 16.84, "grad_norm": 2.6406400203704834, "learning_rate": 0.0001318738539082342, "loss": 1.3959, "step": 573500 }, { "epoch": 16.85, "grad_norm": 1.713273525238037, "learning_rate": 0.0001317267619169028, "loss": 1.3625, "step": 574000 }, { "epoch": 16.87, "grad_norm": 2.819561004638672, "learning_rate": 0.00013157966992557145, "loss": 1.354, "step": 574500 }, { "epoch": 16.88, "grad_norm": 5.471235275268555, "learning_rate": 0.00013143257793424005, "loss": 1.3988, "step": 575000 }, { "epoch": 16.9, "grad_norm": 2.4000468254089355, "learning_rate": 0.0001312854859429087, "loss": 1.3432, "step": 575500 }, { "epoch": 16.91, "grad_norm": 2.053870439529419, "learning_rate": 0.0001311383939515773, "loss": 1.3626, "step": 576000 }, { "epoch": 16.93, "grad_norm": 1.577664852142334, "learning_rate": 0.00013099130196024593, "loss": 1.3806, "step": 576500 }, { "epoch": 16.94, "grad_norm": 16.11113166809082, "learning_rate": 0.00013084420996891454, "loss": 1.3787, "step": 577000 }, { "epoch": 16.96, "grad_norm": 4.257967948913574, "learning_rate": 0.00013069711797758317, "loss": 1.355, "step": 577500 }, { "epoch": 16.97, "grad_norm": 1.8505833148956299, "learning_rate": 0.00013055002598625178, "loss": 1.3702, "step": 578000 }, { "epoch": 16.98, "grad_norm": 2.1093640327453613, "learning_rate": 0.00013040293399492042, "loss": 1.3775, "step": 578500 }, { "epoch": 17.0, "grad_norm": 1.8386383056640625, "learning_rate": 0.00013025584200358902, "loss": 1.3776, "step": 579000 }, { "epoch": 17.01, "grad_norm": 7.34138298034668, "learning_rate": 0.00013010875001225766, "loss": 1.3215, "step": 579500 }, { "epoch": 17.03, "grad_norm": 1.0172581672668457, "learning_rate": 0.00012996165802092627, "loss": 1.3147, "step": 580000 }, { "epoch": 17.04, "grad_norm": 2.6336045265197754, "learning_rate": 0.0001298145660295949, "loss": 1.3365, "step": 580500 }, { "epoch": 17.06, "grad_norm": 3.9906227588653564, "learning_rate": 0.0001296674740382635, "loss": 1.3245, "step": 581000 }, { "epoch": 17.07, "grad_norm": 3.1087284088134766, "learning_rate": 0.00012952038204693214, "loss": 1.2864, "step": 581500 }, { "epoch": 17.09, "grad_norm": 2.5867342948913574, "learning_rate": 0.00012937329005560075, "loss": 1.2947, "step": 582000 }, { "epoch": 17.1, "grad_norm": 1.2118226289749146, "learning_rate": 0.00012922619806426938, "loss": 1.3, "step": 582500 }, { "epoch": 17.12, "grad_norm": 2.56510329246521, "learning_rate": 0.000129079106072938, "loss": 1.3147, "step": 583000 }, { "epoch": 17.13, "grad_norm": 10.831042289733887, "learning_rate": 0.00012893201408160663, "loss": 1.3125, "step": 583500 }, { "epoch": 17.15, "grad_norm": 10.36989688873291, "learning_rate": 0.00012878492209027523, "loss": 1.2969, "step": 584000 }, { "epoch": 17.16, "grad_norm": 2.0238804817199707, "learning_rate": 0.00012863783009894387, "loss": 1.3044, "step": 584500 }, { "epoch": 17.18, "grad_norm": 4.501575469970703, "learning_rate": 0.00012849073810761248, "loss": 1.3013, "step": 585000 }, { "epoch": 17.19, "grad_norm": 4.236315727233887, "learning_rate": 0.0001283436461162811, "loss": 1.2991, "step": 585500 }, { "epoch": 17.21, "grad_norm": 4.139219760894775, "learning_rate": 0.00012819655412494972, "loss": 1.2979, "step": 586000 }, { "epoch": 17.22, "grad_norm": 2.1206071376800537, "learning_rate": 0.00012804946213361835, "loss": 1.3377, "step": 586500 }, { "epoch": 17.23, "grad_norm": 1.7728540897369385, "learning_rate": 0.000127902370142287, "loss": 1.2992, "step": 587000 }, { "epoch": 17.25, "grad_norm": 1.9550994634628296, "learning_rate": 0.0001277552781509556, "loss": 1.3311, "step": 587500 }, { "epoch": 17.26, "grad_norm": 1.6412031650543213, "learning_rate": 0.00012760818615962423, "loss": 1.3101, "step": 588000 }, { "epoch": 17.28, "grad_norm": 1.6908353567123413, "learning_rate": 0.00012746109416829284, "loss": 1.3357, "step": 588500 }, { "epoch": 17.29, "grad_norm": 1.6260554790496826, "learning_rate": 0.00012731400217696147, "loss": 1.3128, "step": 589000 }, { "epoch": 17.31, "grad_norm": 4.529758453369141, "learning_rate": 0.00012716691018563008, "loss": 1.3282, "step": 589500 }, { "epoch": 17.32, "grad_norm": 1.872877836227417, "learning_rate": 0.00012701981819429871, "loss": 1.3271, "step": 590000 }, { "epoch": 17.34, "grad_norm": 3.3399105072021484, "learning_rate": 0.00012687272620296732, "loss": 1.3104, "step": 590500 }, { "epoch": 17.35, "grad_norm": 3.178553819656372, "learning_rate": 0.00012672563421163596, "loss": 1.3222, "step": 591000 }, { "epoch": 17.37, "grad_norm": 2.042067527770996, "learning_rate": 0.00012657854222030456, "loss": 1.332, "step": 591500 }, { "epoch": 17.38, "grad_norm": 2.1601064205169678, "learning_rate": 0.0001264314502289732, "loss": 1.3198, "step": 592000 }, { "epoch": 17.4, "grad_norm": 1.6580477952957153, "learning_rate": 0.0001262843582376418, "loss": 1.3059, "step": 592500 }, { "epoch": 17.41, "grad_norm": 3.175902843475342, "learning_rate": 0.00012613726624631044, "loss": 1.3262, "step": 593000 }, { "epoch": 17.43, "grad_norm": 2.7562525272369385, "learning_rate": 0.00012599017425497905, "loss": 1.3389, "step": 593500 }, { "epoch": 17.44, "grad_norm": 1.7579740285873413, "learning_rate": 0.00012584308226364768, "loss": 1.3308, "step": 594000 }, { "epoch": 17.45, "grad_norm": 4.593905448913574, "learning_rate": 0.0001256959902723163, "loss": 1.3059, "step": 594500 }, { "epoch": 17.47, "grad_norm": 1.438839077949524, "learning_rate": 0.00012554889828098493, "loss": 1.3159, "step": 595000 }, { "epoch": 17.48, "grad_norm": 2.5471925735473633, "learning_rate": 0.00012540180628965353, "loss": 1.2989, "step": 595500 }, { "epoch": 17.5, "grad_norm": 1.8019795417785645, "learning_rate": 0.00012525471429832217, "loss": 1.3441, "step": 596000 }, { "epoch": 17.51, "grad_norm": 2.0826618671417236, "learning_rate": 0.00012510762230699078, "loss": 1.3309, "step": 596500 }, { "epoch": 17.53, "grad_norm": 1.820566177368164, "learning_rate": 0.0001249605303156594, "loss": 1.3224, "step": 597000 }, { "epoch": 17.54, "grad_norm": 24.11446762084961, "learning_rate": 0.00012481343832432802, "loss": 1.2934, "step": 597500 }, { "epoch": 17.56, "grad_norm": 1.5428298711776733, "learning_rate": 0.00012466634633299665, "loss": 1.3277, "step": 598000 }, { "epoch": 17.57, "grad_norm": 2.171504259109497, "learning_rate": 0.00012451925434166526, "loss": 1.3365, "step": 598500 }, { "epoch": 17.59, "grad_norm": 2.862025022506714, "learning_rate": 0.0001243721623503339, "loss": 1.3351, "step": 599000 }, { "epoch": 17.6, "grad_norm": 2.7436113357543945, "learning_rate": 0.0001242250703590025, "loss": 1.3203, "step": 599500 }, { "epoch": 17.62, "grad_norm": 1.9902766942977905, "learning_rate": 0.00012407797836767114, "loss": 1.3324, "step": 600000 }, { "epoch": 17.63, "grad_norm": 3.6419193744659424, "learning_rate": 0.00012393088637633975, "loss": 1.347, "step": 600500 }, { "epoch": 17.65, "grad_norm": 1.8328200578689575, "learning_rate": 0.00012378379438500838, "loss": 1.3261, "step": 601000 }, { "epoch": 17.66, "grad_norm": 3.2816476821899414, "learning_rate": 0.000123636702393677, "loss": 1.3228, "step": 601500 }, { "epoch": 17.67, "grad_norm": 1.142104148864746, "learning_rate": 0.00012348961040234562, "loss": 1.3183, "step": 602000 }, { "epoch": 17.69, "grad_norm": 2.1242527961730957, "learning_rate": 0.00012334251841101423, "loss": 1.3162, "step": 602500 }, { "epoch": 17.7, "grad_norm": 9.647717475891113, "learning_rate": 0.00012319542641968286, "loss": 1.3473, "step": 603000 }, { "epoch": 17.72, "grad_norm": 3.551119804382324, "learning_rate": 0.00012304833442835147, "loss": 1.3513, "step": 603500 }, { "epoch": 17.73, "grad_norm": 1.6903916597366333, "learning_rate": 0.0001229012424370201, "loss": 1.3393, "step": 604000 }, { "epoch": 17.75, "grad_norm": 2.54724383354187, "learning_rate": 0.00012275415044568871, "loss": 1.3402, "step": 604500 }, { "epoch": 17.76, "grad_norm": 5.089727878570557, "learning_rate": 0.00012260705845435735, "loss": 1.303, "step": 605000 }, { "epoch": 17.78, "grad_norm": 12.237126350402832, "learning_rate": 0.00012245996646302596, "loss": 1.3319, "step": 605500 }, { "epoch": 17.79, "grad_norm": 1.576462745666504, "learning_rate": 0.0001223128744716946, "loss": 1.3265, "step": 606000 }, { "epoch": 17.81, "grad_norm": 1.7792675495147705, "learning_rate": 0.0001221657824803632, "loss": 1.3335, "step": 606500 }, { "epoch": 17.82, "grad_norm": 5.532106876373291, "learning_rate": 0.00012201869048903183, "loss": 1.3456, "step": 607000 }, { "epoch": 17.84, "grad_norm": 3.333435297012329, "learning_rate": 0.00012187159849770044, "loss": 1.3332, "step": 607500 }, { "epoch": 17.85, "grad_norm": 3.1190874576568604, "learning_rate": 0.00012172450650636908, "loss": 1.3368, "step": 608000 }, { "epoch": 17.87, "grad_norm": 1.656201720237732, "learning_rate": 0.0001215774145150377, "loss": 1.3266, "step": 608500 }, { "epoch": 17.88, "grad_norm": 2.088050365447998, "learning_rate": 0.00012143032252370632, "loss": 1.3128, "step": 609000 }, { "epoch": 17.9, "grad_norm": 2.941950798034668, "learning_rate": 0.00012128323053237493, "loss": 1.3393, "step": 609500 }, { "epoch": 17.91, "grad_norm": 1.8323218822479248, "learning_rate": 0.00012113613854104356, "loss": 1.3255, "step": 610000 }, { "epoch": 17.92, "grad_norm": 1.3735178709030151, "learning_rate": 0.00012098904654971218, "loss": 1.3095, "step": 610500 }, { "epoch": 17.94, "grad_norm": 2.9366865158081055, "learning_rate": 0.0001208419545583808, "loss": 1.3401, "step": 611000 }, { "epoch": 17.95, "grad_norm": 3.2511463165283203, "learning_rate": 0.00012069486256704941, "loss": 1.3396, "step": 611500 }, { "epoch": 17.97, "grad_norm": 2.2338857650756836, "learning_rate": 0.00012054777057571804, "loss": 1.3286, "step": 612000 }, { "epoch": 17.98, "grad_norm": 1.5889378786087036, "learning_rate": 0.00012040067858438667, "loss": 1.3163, "step": 612500 }, { "epoch": 18.0, "grad_norm": 2.792966365814209, "learning_rate": 0.00012025358659305529, "loss": 1.2973, "step": 613000 }, { "epoch": 18.01, "grad_norm": 2.239032745361328, "learning_rate": 0.00012010649460172391, "loss": 1.3099, "step": 613500 }, { "epoch": 18.03, "grad_norm": 2.293813705444336, "learning_rate": 0.00011995940261039253, "loss": 1.2568, "step": 614000 }, { "epoch": 18.04, "grad_norm": 2.175294876098633, "learning_rate": 0.00011981231061906115, "loss": 1.2661, "step": 614500 }, { "epoch": 18.06, "grad_norm": 1.7673249244689941, "learning_rate": 0.00011966521862772977, "loss": 1.2747, "step": 615000 }, { "epoch": 18.07, "grad_norm": 2.6049957275390625, "learning_rate": 0.00011951812663639839, "loss": 1.2851, "step": 615500 }, { "epoch": 18.09, "grad_norm": 2.0433642864227295, "learning_rate": 0.00011937103464506703, "loss": 1.2846, "step": 616000 }, { "epoch": 18.1, "grad_norm": 2.120561122894287, "learning_rate": 0.00011922394265373563, "loss": 1.2634, "step": 616500 }, { "epoch": 18.12, "grad_norm": 2.6130003929138184, "learning_rate": 0.00011907685066240426, "loss": 1.2844, "step": 617000 }, { "epoch": 18.13, "grad_norm": 1.8312240839004517, "learning_rate": 0.00011892975867107288, "loss": 1.289, "step": 617500 }, { "epoch": 18.14, "grad_norm": 1.5402841567993164, "learning_rate": 0.00011878266667974151, "loss": 1.2578, "step": 618000 }, { "epoch": 18.16, "grad_norm": 2.0804052352905273, "learning_rate": 0.00011863557468841012, "loss": 1.2558, "step": 618500 }, { "epoch": 18.17, "grad_norm": 1.77811861038208, "learning_rate": 0.00011848848269707874, "loss": 1.2853, "step": 619000 }, { "epoch": 18.19, "grad_norm": 6.067068099975586, "learning_rate": 0.00011834139070574736, "loss": 1.2941, "step": 619500 }, { "epoch": 18.2, "grad_norm": 2.4684622287750244, "learning_rate": 0.000118194298714416, "loss": 1.2749, "step": 620000 }, { "epoch": 18.22, "grad_norm": 3.7913448810577393, "learning_rate": 0.0001180472067230846, "loss": 1.298, "step": 620500 }, { "epoch": 18.23, "grad_norm": 1.343802571296692, "learning_rate": 0.00011790011473175322, "loss": 1.2954, "step": 621000 }, { "epoch": 18.25, "grad_norm": 1.797194480895996, "learning_rate": 0.00011775302274042185, "loss": 1.2718, "step": 621500 }, { "epoch": 18.26, "grad_norm": 2.3011558055877686, "learning_rate": 0.00011760593074909048, "loss": 1.2966, "step": 622000 }, { "epoch": 18.28, "grad_norm": 2.5689167976379395, "learning_rate": 0.00011745883875775909, "loss": 1.286, "step": 622500 }, { "epoch": 18.29, "grad_norm": 8.428597450256348, "learning_rate": 0.00011731174676642771, "loss": 1.2872, "step": 623000 }, { "epoch": 18.31, "grad_norm": 1.1936590671539307, "learning_rate": 0.00011716465477509633, "loss": 1.2621, "step": 623500 }, { "epoch": 18.32, "grad_norm": 1.5052251815795898, "learning_rate": 0.00011701756278376497, "loss": 1.2758, "step": 624000 }, { "epoch": 18.34, "grad_norm": 1.296823263168335, "learning_rate": 0.00011687047079243357, "loss": 1.266, "step": 624500 }, { "epoch": 18.35, "grad_norm": 3.121631383895874, "learning_rate": 0.0001167233788011022, "loss": 1.2858, "step": 625000 }, { "epoch": 18.36, "grad_norm": 7.022789478302002, "learning_rate": 0.00011657628680977081, "loss": 1.2986, "step": 625500 }, { "epoch": 18.38, "grad_norm": 11.550426483154297, "learning_rate": 0.00011642919481843945, "loss": 1.2736, "step": 626000 }, { "epoch": 18.39, "grad_norm": 2.277326822280884, "learning_rate": 0.00011628210282710807, "loss": 1.3052, "step": 626500 }, { "epoch": 18.41, "grad_norm": 3.439568519592285, "learning_rate": 0.00011613501083577668, "loss": 1.3042, "step": 627000 }, { "epoch": 18.42, "grad_norm": 8.868010520935059, "learning_rate": 0.0001159879188444453, "loss": 1.2612, "step": 627500 }, { "epoch": 18.44, "grad_norm": 1.5216681957244873, "learning_rate": 0.00011584082685311393, "loss": 1.2634, "step": 628000 }, { "epoch": 18.45, "grad_norm": 2.113112688064575, "learning_rate": 0.00011569373486178256, "loss": 1.2681, "step": 628500 }, { "epoch": 18.47, "grad_norm": 2.9850034713745117, "learning_rate": 0.00011554664287045116, "loss": 1.2865, "step": 629000 }, { "epoch": 18.48, "grad_norm": 2.309042453765869, "learning_rate": 0.00011539955087911978, "loss": 1.2674, "step": 629500 }, { "epoch": 18.5, "grad_norm": 3.1562564373016357, "learning_rate": 0.00011525245888778842, "loss": 1.262, "step": 630000 }, { "epoch": 18.51, "grad_norm": 3.4192593097686768, "learning_rate": 0.00011510536689645704, "loss": 1.2729, "step": 630500 }, { "epoch": 18.53, "grad_norm": 2.193237543106079, "learning_rate": 0.00011495827490512565, "loss": 1.2586, "step": 631000 }, { "epoch": 18.54, "grad_norm": 8.091324806213379, "learning_rate": 0.00011481118291379427, "loss": 1.2649, "step": 631500 }, { "epoch": 18.56, "grad_norm": 2.0321621894836426, "learning_rate": 0.0001146640909224629, "loss": 1.2792, "step": 632000 }, { "epoch": 18.57, "grad_norm": 1.623028039932251, "learning_rate": 0.00011451699893113152, "loss": 1.3013, "step": 632500 }, { "epoch": 18.59, "grad_norm": 2.7462871074676514, "learning_rate": 0.00011436990693980013, "loss": 1.2858, "step": 633000 }, { "epoch": 18.6, "grad_norm": 22.743488311767578, "learning_rate": 0.00011422281494846875, "loss": 1.2739, "step": 633500 }, { "epoch": 18.61, "grad_norm": 4.659852981567383, "learning_rate": 0.00011407572295713739, "loss": 1.2773, "step": 634000 }, { "epoch": 18.63, "grad_norm": 1.8134876489639282, "learning_rate": 0.00011392863096580601, "loss": 1.274, "step": 634500 }, { "epoch": 18.64, "grad_norm": 2.266272783279419, "learning_rate": 0.00011378153897447463, "loss": 1.3084, "step": 635000 }, { "epoch": 18.66, "grad_norm": 1.4625264406204224, "learning_rate": 0.00011363444698314324, "loss": 1.2794, "step": 635500 }, { "epoch": 18.67, "grad_norm": 1.57483971118927, "learning_rate": 0.00011348735499181187, "loss": 1.2872, "step": 636000 }, { "epoch": 18.69, "grad_norm": 1.3935645818710327, "learning_rate": 0.00011334026300048049, "loss": 1.2809, "step": 636500 }, { "epoch": 18.7, "grad_norm": 3.3207247257232666, "learning_rate": 0.00011319317100914911, "loss": 1.2764, "step": 637000 }, { "epoch": 18.72, "grad_norm": 4.18394660949707, "learning_rate": 0.00011304607901781772, "loss": 1.2855, "step": 637500 }, { "epoch": 18.73, "grad_norm": 1.6960937976837158, "learning_rate": 0.00011289898702648636, "loss": 1.2807, "step": 638000 }, { "epoch": 18.75, "grad_norm": 6.864727020263672, "learning_rate": 0.00011275189503515498, "loss": 1.2736, "step": 638500 }, { "epoch": 18.76, "grad_norm": 1.7619363069534302, "learning_rate": 0.0001126048030438236, "loss": 1.3063, "step": 639000 }, { "epoch": 18.78, "grad_norm": 1.6224156618118286, "learning_rate": 0.0001124577110524922, "loss": 1.3031, "step": 639500 }, { "epoch": 18.79, "grad_norm": 4.564239025115967, "learning_rate": 0.00011231061906116084, "loss": 1.3185, "step": 640000 }, { "epoch": 18.81, "grad_norm": 2.237443208694458, "learning_rate": 0.00011216352706982946, "loss": 1.2612, "step": 640500 }, { "epoch": 18.82, "grad_norm": 1.7501612901687622, "learning_rate": 0.00011201643507849808, "loss": 1.278, "step": 641000 }, { "epoch": 18.83, "grad_norm": 1.1969166994094849, "learning_rate": 0.00011186934308716669, "loss": 1.2979, "step": 641500 }, { "epoch": 18.85, "grad_norm": 2.23476505279541, "learning_rate": 0.00011172225109583533, "loss": 1.2796, "step": 642000 }, { "epoch": 18.86, "grad_norm": 3.1486740112304688, "learning_rate": 0.00011157515910450395, "loss": 1.3212, "step": 642500 }, { "epoch": 18.88, "grad_norm": 3.5119261741638184, "learning_rate": 0.00011142806711317257, "loss": 1.3082, "step": 643000 }, { "epoch": 18.89, "grad_norm": 3.0694100856781006, "learning_rate": 0.00011128097512184118, "loss": 1.2643, "step": 643500 }, { "epoch": 18.91, "grad_norm": 2.235492467880249, "learning_rate": 0.00011113388313050981, "loss": 1.2832, "step": 644000 }, { "epoch": 18.92, "grad_norm": 2.627898931503296, "learning_rate": 0.00011098679113917843, "loss": 1.2863, "step": 644500 }, { "epoch": 18.94, "grad_norm": 1.390758991241455, "learning_rate": 0.00011083969914784705, "loss": 1.2949, "step": 645000 }, { "epoch": 18.95, "grad_norm": 1.9412230253219604, "learning_rate": 0.00011069260715651567, "loss": 1.2765, "step": 645500 }, { "epoch": 18.97, "grad_norm": 7.697941780090332, "learning_rate": 0.0001105455151651843, "loss": 1.2872, "step": 646000 }, { "epoch": 18.98, "grad_norm": 2.0770368576049805, "learning_rate": 0.00011039842317385292, "loss": 1.3058, "step": 646500 }, { "epoch": 19.0, "grad_norm": 1.8311492204666138, "learning_rate": 0.00011025133118252154, "loss": 1.3229, "step": 647000 }, { "epoch": 19.01, "grad_norm": 2.744004726409912, "learning_rate": 0.00011010423919119017, "loss": 1.253, "step": 647500 }, { "epoch": 19.03, "grad_norm": 2.0883147716522217, "learning_rate": 0.00010995714719985879, "loss": 1.2348, "step": 648000 }, { "epoch": 19.04, "grad_norm": 13.648962020874023, "learning_rate": 0.0001098100552085274, "loss": 1.2572, "step": 648500 }, { "epoch": 19.05, "grad_norm": 3.89188814163208, "learning_rate": 0.00010966296321719602, "loss": 1.2312, "step": 649000 }, { "epoch": 19.07, "grad_norm": 12.498522758483887, "learning_rate": 0.00010951587122586466, "loss": 1.2285, "step": 649500 }, { "epoch": 19.08, "grad_norm": 1.9123058319091797, "learning_rate": 0.00010936877923453328, "loss": 1.2223, "step": 650000 }, { "epoch": 19.1, "grad_norm": 1.9629552364349365, "learning_rate": 0.00010922168724320188, "loss": 1.2323, "step": 650500 }, { "epoch": 19.11, "grad_norm": 16.135618209838867, "learning_rate": 0.0001090745952518705, "loss": 1.1976, "step": 651000 }, { "epoch": 19.13, "grad_norm": 2.948089361190796, "learning_rate": 0.00010892750326053914, "loss": 1.2367, "step": 651500 }, { "epoch": 19.14, "grad_norm": 2.4549195766448975, "learning_rate": 0.00010878041126920776, "loss": 1.2345, "step": 652000 }, { "epoch": 19.16, "grad_norm": 1.7298622131347656, "learning_rate": 0.00010863331927787637, "loss": 1.2477, "step": 652500 }, { "epoch": 19.17, "grad_norm": 3.237170696258545, "learning_rate": 0.00010848622728654499, "loss": 1.2578, "step": 653000 }, { "epoch": 19.19, "grad_norm": 2.876091718673706, "learning_rate": 0.00010833913529521362, "loss": 1.2657, "step": 653500 }, { "epoch": 19.2, "grad_norm": 2.6806657314300537, "learning_rate": 0.00010819204330388225, "loss": 1.2369, "step": 654000 }, { "epoch": 19.22, "grad_norm": 1.682861328125, "learning_rate": 0.00010804495131255085, "loss": 1.2353, "step": 654500 }, { "epoch": 19.23, "grad_norm": 1.420599102973938, "learning_rate": 0.00010789785932121947, "loss": 1.2292, "step": 655000 }, { "epoch": 19.25, "grad_norm": 2.785423517227173, "learning_rate": 0.00010775076732988811, "loss": 1.2299, "step": 655500 }, { "epoch": 19.26, "grad_norm": 3.515298843383789, "learning_rate": 0.00010760367533855673, "loss": 1.2343, "step": 656000 }, { "epoch": 19.28, "grad_norm": 8.15224552154541, "learning_rate": 0.00010745658334722534, "loss": 1.262, "step": 656500 }, { "epoch": 19.29, "grad_norm": 2.5358471870422363, "learning_rate": 0.00010730949135589396, "loss": 1.2579, "step": 657000 }, { "epoch": 19.3, "grad_norm": 2.8267860412597656, "learning_rate": 0.0001071623993645626, "loss": 1.2195, "step": 657500 }, { "epoch": 19.32, "grad_norm": 2.0857648849487305, "learning_rate": 0.00010701530737323122, "loss": 1.2244, "step": 658000 }, { "epoch": 19.33, "grad_norm": 2.2825379371643066, "learning_rate": 0.00010686821538189984, "loss": 1.2502, "step": 658500 }, { "epoch": 19.35, "grad_norm": 2.1249475479125977, "learning_rate": 0.00010672112339056844, "loss": 1.2408, "step": 659000 }, { "epoch": 19.36, "grad_norm": 1.9578863382339478, "learning_rate": 0.00010657403139923708, "loss": 1.2599, "step": 659500 }, { "epoch": 19.38, "grad_norm": 1.7473647594451904, "learning_rate": 0.0001064269394079057, "loss": 1.2482, "step": 660000 }, { "epoch": 19.39, "grad_norm": 1.9577364921569824, "learning_rate": 0.00010627984741657432, "loss": 1.2389, "step": 660500 }, { "epoch": 19.41, "grad_norm": 6.437145233154297, "learning_rate": 0.00010613275542524293, "loss": 1.2327, "step": 661000 }, { "epoch": 19.42, "grad_norm": 36.84048080444336, "learning_rate": 0.00010598566343391156, "loss": 1.2553, "step": 661500 }, { "epoch": 19.44, "grad_norm": 2.4691734313964844, "learning_rate": 0.00010583857144258018, "loss": 1.2551, "step": 662000 }, { "epoch": 19.45, "grad_norm": 3.6141631603240967, "learning_rate": 0.0001056914794512488, "loss": 1.2459, "step": 662500 }, { "epoch": 19.47, "grad_norm": 4.1790008544921875, "learning_rate": 0.00010554438745991741, "loss": 1.2299, "step": 663000 }, { "epoch": 19.48, "grad_norm": 2.3077800273895264, "learning_rate": 0.00010539729546858605, "loss": 1.2354, "step": 663500 }, { "epoch": 19.5, "grad_norm": 2.0375678539276123, "learning_rate": 0.00010525020347725467, "loss": 1.2332, "step": 664000 }, { "epoch": 19.51, "grad_norm": 2.190852403640747, "learning_rate": 0.00010510311148592329, "loss": 1.2343, "step": 664500 }, { "epoch": 19.52, "grad_norm": 1.7503656148910522, "learning_rate": 0.0001049560194945919, "loss": 1.2328, "step": 665000 }, { "epoch": 19.54, "grad_norm": 2.364180326461792, "learning_rate": 0.00010480892750326053, "loss": 1.245, "step": 665500 }, { "epoch": 19.55, "grad_norm": 6.837544918060303, "learning_rate": 0.00010466183551192915, "loss": 1.2361, "step": 666000 }, { "epoch": 19.57, "grad_norm": 18.344188690185547, "learning_rate": 0.00010451474352059777, "loss": 1.2527, "step": 666500 }, { "epoch": 19.58, "grad_norm": 4.090867519378662, "learning_rate": 0.00010436765152926638, "loss": 1.2573, "step": 667000 }, { "epoch": 19.6, "grad_norm": 61.80951690673828, "learning_rate": 0.00010422055953793502, "loss": 1.2776, "step": 667500 }, { "epoch": 19.61, "grad_norm": 1.9649507999420166, "learning_rate": 0.00010407346754660364, "loss": 1.2479, "step": 668000 }, { "epoch": 19.63, "grad_norm": 4.030837535858154, "learning_rate": 0.00010392637555527226, "loss": 1.2458, "step": 668500 }, { "epoch": 19.64, "grad_norm": 3.310805082321167, "learning_rate": 0.00010377928356394088, "loss": 1.2491, "step": 669000 }, { "epoch": 19.66, "grad_norm": 6.558318138122559, "learning_rate": 0.0001036321915726095, "loss": 1.2562, "step": 669500 }, { "epoch": 19.67, "grad_norm": 2.0995540618896484, "learning_rate": 0.00010348509958127812, "loss": 1.2529, "step": 670000 }, { "epoch": 19.69, "grad_norm": 3.003690242767334, "learning_rate": 0.00010333800758994674, "loss": 1.2717, "step": 670500 }, { "epoch": 19.7, "grad_norm": 2.099637269973755, "learning_rate": 0.00010319091559861536, "loss": 1.2632, "step": 671000 }, { "epoch": 19.72, "grad_norm": 35.86410140991211, "learning_rate": 0.000103043823607284, "loss": 1.2763, "step": 671500 }, { "epoch": 19.73, "grad_norm": 1.7510465383529663, "learning_rate": 0.0001028967316159526, "loss": 1.2535, "step": 672000 }, { "epoch": 19.74, "grad_norm": 10.661267280578613, "learning_rate": 0.00010274963962462123, "loss": 1.2346, "step": 672500 }, { "epoch": 19.76, "grad_norm": 1.9645477533340454, "learning_rate": 0.00010260254763328985, "loss": 1.2398, "step": 673000 }, { "epoch": 19.77, "grad_norm": 2.795703172683716, "learning_rate": 0.00010245545564195848, "loss": 1.2683, "step": 673500 }, { "epoch": 19.79, "grad_norm": 3.7908451557159424, "learning_rate": 0.00010230836365062709, "loss": 1.2439, "step": 674000 }, { "epoch": 19.8, "grad_norm": 2.028703451156616, "learning_rate": 0.00010216127165929571, "loss": 1.2518, "step": 674500 }, { "epoch": 19.82, "grad_norm": 2.059154510498047, "learning_rate": 0.00010201417966796433, "loss": 1.2359, "step": 675000 }, { "epoch": 19.83, "grad_norm": 1.3817317485809326, "learning_rate": 0.00010186708767663297, "loss": 1.27, "step": 675500 }, { "epoch": 19.85, "grad_norm": 1.8642240762710571, "learning_rate": 0.00010171999568530158, "loss": 1.2604, "step": 676000 }, { "epoch": 19.86, "grad_norm": 1.7178900241851807, "learning_rate": 0.0001015729036939702, "loss": 1.2509, "step": 676500 }, { "epoch": 19.88, "grad_norm": 6.73402738571167, "learning_rate": 0.00010142581170263882, "loss": 1.2566, "step": 677000 }, { "epoch": 19.89, "grad_norm": 1.7260433435440063, "learning_rate": 0.00010127871971130745, "loss": 1.258, "step": 677500 }, { "epoch": 19.91, "grad_norm": 2.0348527431488037, "learning_rate": 0.00010113162771997606, "loss": 1.234, "step": 678000 }, { "epoch": 19.92, "grad_norm": 3.1736955642700195, "learning_rate": 0.00010098453572864468, "loss": 1.241, "step": 678500 }, { "epoch": 19.94, "grad_norm": 10.23302173614502, "learning_rate": 0.0001008374437373133, "loss": 1.2604, "step": 679000 }, { "epoch": 19.95, "grad_norm": 2.974153995513916, "learning_rate": 0.00010069035174598194, "loss": 1.2525, "step": 679500 }, { "epoch": 19.97, "grad_norm": 2.5226101875305176, "learning_rate": 0.00010054325975465056, "loss": 1.264, "step": 680000 }, { "epoch": 19.98, "grad_norm": 2.472259521484375, "learning_rate": 0.00010039616776331917, "loss": 1.2299, "step": 680500 }, { "epoch": 19.99, "grad_norm": 2.3573238849639893, "learning_rate": 0.00010024907577198779, "loss": 1.2657, "step": 681000 }, { "epoch": 20.01, "grad_norm": 3.6109812259674072, "learning_rate": 0.00010010198378065642, "loss": 1.2247, "step": 681500 }, { "epoch": 20.02, "grad_norm": 1.8781336545944214, "learning_rate": 9.995489178932504e-05, "loss": 1.2068, "step": 682000 }, { "epoch": 20.04, "grad_norm": 1.4420850276947021, "learning_rate": 9.980779979799365e-05, "loss": 1.2, "step": 682500 }, { "epoch": 20.05, "grad_norm": 3.082235097885132, "learning_rate": 9.966070780666227e-05, "loss": 1.2062, "step": 683000 }, { "epoch": 20.07, "grad_norm": 1.9524058103561401, "learning_rate": 9.95136158153309e-05, "loss": 1.1999, "step": 683500 }, { "epoch": 20.08, "grad_norm": 1.9696966409683228, "learning_rate": 9.936652382399953e-05, "loss": 1.1944, "step": 684000 }, { "epoch": 20.1, "grad_norm": 3.851034641265869, "learning_rate": 9.921943183266813e-05, "loss": 1.1913, "step": 684500 }, { "epoch": 20.11, "grad_norm": 1.7799595594406128, "learning_rate": 9.907233984133676e-05, "loss": 1.21, "step": 685000 }, { "epoch": 20.13, "grad_norm": 2.3180084228515625, "learning_rate": 9.892524785000539e-05, "loss": 1.1741, "step": 685500 }, { "epoch": 20.14, "grad_norm": 2.012601375579834, "learning_rate": 9.877815585867401e-05, "loss": 1.1956, "step": 686000 }, { "epoch": 20.16, "grad_norm": 2.5793349742889404, "learning_rate": 9.863106386734262e-05, "loss": 1.2011, "step": 686500 }, { "epoch": 20.17, "grad_norm": 1.7113804817199707, "learning_rate": 9.848397187601124e-05, "loss": 1.2382, "step": 687000 }, { "epoch": 20.19, "grad_norm": 1.8239011764526367, "learning_rate": 9.833687988467988e-05, "loss": 1.1904, "step": 687500 }, { "epoch": 20.2, "grad_norm": 8.440109252929688, "learning_rate": 9.81897878933485e-05, "loss": 1.2087, "step": 688000 }, { "epoch": 20.21, "grad_norm": 2.4333038330078125, "learning_rate": 9.80426959020171e-05, "loss": 1.209, "step": 688500 }, { "epoch": 20.23, "grad_norm": 3.217632532119751, "learning_rate": 9.789560391068572e-05, "loss": 1.1964, "step": 689000 }, { "epoch": 20.24, "grad_norm": 3.0806314945220947, "learning_rate": 9.774851191935436e-05, "loss": 1.2227, "step": 689500 }, { "epoch": 20.26, "grad_norm": 21.035554885864258, "learning_rate": 9.760141992802298e-05, "loss": 1.1899, "step": 690000 }, { "epoch": 20.27, "grad_norm": 4.15212345123291, "learning_rate": 9.74543279366916e-05, "loss": 1.1919, "step": 690500 }, { "epoch": 20.29, "grad_norm": 8.08483600616455, "learning_rate": 9.730723594536021e-05, "loss": 1.2083, "step": 691000 }, { "epoch": 20.3, "grad_norm": 2.5341944694519043, "learning_rate": 9.716014395402884e-05, "loss": 1.181, "step": 691500 }, { "epoch": 20.32, "grad_norm": 7.360681533813477, "learning_rate": 9.701305196269747e-05, "loss": 1.1975, "step": 692000 }, { "epoch": 20.33, "grad_norm": 1.5373992919921875, "learning_rate": 9.686595997136609e-05, "loss": 1.191, "step": 692500 }, { "epoch": 20.35, "grad_norm": 1.8679372072219849, "learning_rate": 9.67188679800347e-05, "loss": 1.213, "step": 693000 }, { "epoch": 20.36, "grad_norm": 3.206306219100952, "learning_rate": 9.657177598870333e-05, "loss": 1.2051, "step": 693500 }, { "epoch": 20.38, "grad_norm": 2.6320650577545166, "learning_rate": 9.642468399737195e-05, "loss": 1.1851, "step": 694000 }, { "epoch": 20.39, "grad_norm": 3.6654086112976074, "learning_rate": 9.627759200604057e-05, "loss": 1.2063, "step": 694500 }, { "epoch": 20.41, "grad_norm": 3.010348081588745, "learning_rate": 9.613050001470918e-05, "loss": 1.2239, "step": 695000 }, { "epoch": 20.42, "grad_norm": 27.742935180664062, "learning_rate": 9.598340802337781e-05, "loss": 1.2084, "step": 695500 }, { "epoch": 20.43, "grad_norm": 2.783250570297241, "learning_rate": 9.583631603204643e-05, "loss": 1.2263, "step": 696000 }, { "epoch": 20.45, "grad_norm": 2.516063690185547, "learning_rate": 9.568922404071506e-05, "loss": 1.2084, "step": 696500 }, { "epoch": 20.46, "grad_norm": 1.8266417980194092, "learning_rate": 9.554213204938366e-05, "loss": 1.2052, "step": 697000 }, { "epoch": 20.48, "grad_norm": 1.2132940292358398, "learning_rate": 9.53950400580523e-05, "loss": 1.2132, "step": 697500 }, { "epoch": 20.49, "grad_norm": 4.550230503082275, "learning_rate": 9.524794806672092e-05, "loss": 1.2263, "step": 698000 }, { "epoch": 20.51, "grad_norm": 2.1874680519104004, "learning_rate": 9.510085607538954e-05, "loss": 1.1855, "step": 698500 }, { "epoch": 20.52, "grad_norm": 1.8161512613296509, "learning_rate": 9.495376408405815e-05, "loss": 1.1949, "step": 699000 }, { "epoch": 20.54, "grad_norm": 1.3943161964416504, "learning_rate": 9.480667209272678e-05, "loss": 1.1944, "step": 699500 }, { "epoch": 20.55, "grad_norm": 3.2997055053710938, "learning_rate": 9.46595801013954e-05, "loss": 1.2213, "step": 700000 }, { "epoch": 20.57, "grad_norm": 1.9309003353118896, "learning_rate": 9.451248811006402e-05, "loss": 1.2163, "step": 700500 }, { "epoch": 20.58, "grad_norm": 2.734384775161743, "learning_rate": 9.436539611873265e-05, "loss": 1.1912, "step": 701000 }, { "epoch": 20.6, "grad_norm": 2.1459152698516846, "learning_rate": 9.421830412740127e-05, "loss": 1.2343, "step": 701500 }, { "epoch": 20.61, "grad_norm": 2.039071798324585, "learning_rate": 9.407121213606989e-05, "loss": 1.228, "step": 702000 }, { "epoch": 20.63, "grad_norm": 3.1236155033111572, "learning_rate": 9.392412014473851e-05, "loss": 1.2362, "step": 702500 }, { "epoch": 20.64, "grad_norm": 6.471434593200684, "learning_rate": 9.377702815340713e-05, "loss": 1.2159, "step": 703000 }, { "epoch": 20.66, "grad_norm": 3.2641398906707764, "learning_rate": 9.362993616207576e-05, "loss": 1.1702, "step": 703500 }, { "epoch": 20.67, "grad_norm": 2.0481953620910645, "learning_rate": 9.348284417074437e-05, "loss": 1.218, "step": 704000 }, { "epoch": 20.68, "grad_norm": 1.6071834564208984, "learning_rate": 9.3335752179413e-05, "loss": 1.2422, "step": 704500 }, { "epoch": 20.7, "grad_norm": 1.8020005226135254, "learning_rate": 9.318866018808163e-05, "loss": 1.2268, "step": 705000 }, { "epoch": 20.71, "grad_norm": 4.970008373260498, "learning_rate": 9.304156819675025e-05, "loss": 1.2093, "step": 705500 }, { "epoch": 20.73, "grad_norm": 2.152195453643799, "learning_rate": 9.289447620541886e-05, "loss": 1.2361, "step": 706000 }, { "epoch": 20.74, "grad_norm": 5.956038475036621, "learning_rate": 9.274738421408748e-05, "loss": 1.1908, "step": 706500 }, { "epoch": 20.76, "grad_norm": 2.429544687271118, "learning_rate": 9.260029222275611e-05, "loss": 1.1805, "step": 707000 }, { "epoch": 20.77, "grad_norm": 2.6851959228515625, "learning_rate": 9.245320023142473e-05, "loss": 1.204, "step": 707500 }, { "epoch": 20.79, "grad_norm": 2.6581480503082275, "learning_rate": 9.230610824009334e-05, "loss": 1.2357, "step": 708000 }, { "epoch": 20.8, "grad_norm": 2.938990354537964, "learning_rate": 9.215901624876196e-05, "loss": 1.2326, "step": 708500 }, { "epoch": 20.82, "grad_norm": 1.9101125001907349, "learning_rate": 9.20119242574306e-05, "loss": 1.2013, "step": 709000 }, { "epoch": 20.83, "grad_norm": 3.4203646183013916, "learning_rate": 9.186483226609922e-05, "loss": 1.2021, "step": 709500 }, { "epoch": 20.85, "grad_norm": 2.2764744758605957, "learning_rate": 9.171774027476783e-05, "loss": 1.223, "step": 710000 }, { "epoch": 20.86, "grad_norm": 1.5339056253433228, "learning_rate": 9.157064828343645e-05, "loss": 1.2001, "step": 710500 }, { "epoch": 20.88, "grad_norm": 1.3566075563430786, "learning_rate": 9.142355629210508e-05, "loss": 1.2324, "step": 711000 }, { "epoch": 20.89, "grad_norm": 1.5568170547485352, "learning_rate": 9.12764643007737e-05, "loss": 1.2065, "step": 711500 }, { "epoch": 20.9, "grad_norm": 26.298019409179688, "learning_rate": 9.112937230944231e-05, "loss": 1.2283, "step": 712000 }, { "epoch": 20.92, "grad_norm": 1.2808449268341064, "learning_rate": 9.098228031811093e-05, "loss": 1.2531, "step": 712500 }, { "epoch": 20.93, "grad_norm": 1.9005004167556763, "learning_rate": 9.083518832677957e-05, "loss": 1.2108, "step": 713000 }, { "epoch": 20.95, "grad_norm": 6.3783721923828125, "learning_rate": 9.068809633544819e-05, "loss": 1.2015, "step": 713500 }, { "epoch": 20.96, "grad_norm": 2.5035910606384277, "learning_rate": 9.054100434411681e-05, "loss": 1.2342, "step": 714000 }, { "epoch": 20.98, "grad_norm": 9.022185325622559, "learning_rate": 9.039391235278542e-05, "loss": 1.2176, "step": 714500 }, { "epoch": 20.99, "grad_norm": 1.442610502243042, "learning_rate": 9.024682036145405e-05, "loss": 1.2406, "step": 715000 }, { "epoch": 21.01, "grad_norm": 2.419548273086548, "learning_rate": 9.009972837012267e-05, "loss": 1.1987, "step": 715500 }, { "epoch": 21.02, "grad_norm": 9.644920349121094, "learning_rate": 8.995263637879129e-05, "loss": 1.141, "step": 716000 }, { "epoch": 21.04, "grad_norm": 2.2750537395477295, "learning_rate": 8.98055443874599e-05, "loss": 1.1621, "step": 716500 }, { "epoch": 21.05, "grad_norm": 12.630777359008789, "learning_rate": 8.965845239612853e-05, "loss": 1.1495, "step": 717000 }, { "epoch": 21.07, "grad_norm": 1.7819632291793823, "learning_rate": 8.951136040479716e-05, "loss": 1.1687, "step": 717500 }, { "epoch": 21.08, "grad_norm": 2.4531936645507812, "learning_rate": 8.936426841346578e-05, "loss": 1.1722, "step": 718000 }, { "epoch": 21.1, "grad_norm": 4.26564884185791, "learning_rate": 8.921717642213438e-05, "loss": 1.1804, "step": 718500 }, { "epoch": 21.11, "grad_norm": 3.8713653087615967, "learning_rate": 8.907008443080302e-05, "loss": 1.1954, "step": 719000 }, { "epoch": 21.12, "grad_norm": 2.3605353832244873, "learning_rate": 8.892299243947164e-05, "loss": 1.1688, "step": 719500 }, { "epoch": 21.14, "grad_norm": 2.322021007537842, "learning_rate": 8.877590044814026e-05, "loss": 1.1403, "step": 720000 }, { "epoch": 21.15, "grad_norm": 2.5785558223724365, "learning_rate": 8.862880845680887e-05, "loss": 1.1812, "step": 720500 }, { "epoch": 21.17, "grad_norm": 1.8795828819274902, "learning_rate": 8.84817164654775e-05, "loss": 1.1756, "step": 721000 }, { "epoch": 21.18, "grad_norm": 1.7690235376358032, "learning_rate": 8.833462447414613e-05, "loss": 1.1592, "step": 721500 }, { "epoch": 21.2, "grad_norm": 2.3973443508148193, "learning_rate": 8.818753248281475e-05, "loss": 1.1586, "step": 722000 }, { "epoch": 21.21, "grad_norm": 11.784666061401367, "learning_rate": 8.804044049148335e-05, "loss": 1.1522, "step": 722500 }, { "epoch": 21.23, "grad_norm": 1.8928039073944092, "learning_rate": 8.789334850015199e-05, "loss": 1.1739, "step": 723000 }, { "epoch": 21.24, "grad_norm": 28.14616584777832, "learning_rate": 8.774625650882061e-05, "loss": 1.1844, "step": 723500 }, { "epoch": 21.26, "grad_norm": 22.186017990112305, "learning_rate": 8.759916451748923e-05, "loss": 1.1501, "step": 724000 }, { "epoch": 21.27, "grad_norm": 3.089933395385742, "learning_rate": 8.745207252615785e-05, "loss": 1.1687, "step": 724500 }, { "epoch": 21.29, "grad_norm": 3.315870523452759, "learning_rate": 8.730498053482647e-05, "loss": 1.1695, "step": 725000 }, { "epoch": 21.3, "grad_norm": 2.118852138519287, "learning_rate": 8.71578885434951e-05, "loss": 1.1831, "step": 725500 }, { "epoch": 21.32, "grad_norm": 2.3328022956848145, "learning_rate": 8.701079655216372e-05, "loss": 1.1972, "step": 726000 }, { "epoch": 21.33, "grad_norm": 2.6026008129119873, "learning_rate": 8.686370456083234e-05, "loss": 1.177, "step": 726500 }, { "epoch": 21.34, "grad_norm": 1.873986840248108, "learning_rate": 8.671661256950097e-05, "loss": 1.1856, "step": 727000 }, { "epoch": 21.36, "grad_norm": 2.4924209117889404, "learning_rate": 8.656952057816958e-05, "loss": 1.1587, "step": 727500 }, { "epoch": 21.37, "grad_norm": 3.228402614593506, "learning_rate": 8.64224285868382e-05, "loss": 1.1654, "step": 728000 }, { "epoch": 21.39, "grad_norm": 2.2008731365203857, "learning_rate": 8.627533659550682e-05, "loss": 1.1953, "step": 728500 }, { "epoch": 21.4, "grad_norm": 16.22712516784668, "learning_rate": 8.612824460417546e-05, "loss": 1.1871, "step": 729000 }, { "epoch": 21.42, "grad_norm": 4.7584919929504395, "learning_rate": 8.598115261284406e-05, "loss": 1.1865, "step": 729500 }, { "epoch": 21.43, "grad_norm": 3.2225780487060547, "learning_rate": 8.583406062151268e-05, "loss": 1.1776, "step": 730000 }, { "epoch": 21.45, "grad_norm": 5.469396591186523, "learning_rate": 8.56869686301813e-05, "loss": 1.1897, "step": 730500 }, { "epoch": 21.46, "grad_norm": 1.8942010402679443, "learning_rate": 8.553987663884994e-05, "loss": 1.1484, "step": 731000 }, { "epoch": 21.48, "grad_norm": 2.122103452682495, "learning_rate": 8.539278464751855e-05, "loss": 1.1707, "step": 731500 }, { "epoch": 21.49, "grad_norm": 3.3333709239959717, "learning_rate": 8.524569265618717e-05, "loss": 1.1471, "step": 732000 }, { "epoch": 21.51, "grad_norm": 3.002875566482544, "learning_rate": 8.509860066485579e-05, "loss": 1.1785, "step": 732500 }, { "epoch": 21.52, "grad_norm": 3.3468096256256104, "learning_rate": 8.495150867352442e-05, "loss": 1.1705, "step": 733000 }, { "epoch": 21.54, "grad_norm": 20.700353622436523, "learning_rate": 8.480441668219303e-05, "loss": 1.1792, "step": 733500 }, { "epoch": 21.55, "grad_norm": 5.712481498718262, "learning_rate": 8.465732469086165e-05, "loss": 1.1633, "step": 734000 }, { "epoch": 21.57, "grad_norm": 2.2254440784454346, "learning_rate": 8.451023269953027e-05, "loss": 1.1809, "step": 734500 }, { "epoch": 21.58, "grad_norm": 5.889398574829102, "learning_rate": 8.436314070819891e-05, "loss": 1.1732, "step": 735000 }, { "epoch": 21.59, "grad_norm": 6.2943572998046875, "learning_rate": 8.421604871686753e-05, "loss": 1.1896, "step": 735500 }, { "epoch": 21.61, "grad_norm": 2.1915676593780518, "learning_rate": 8.406895672553614e-05, "loss": 1.1884, "step": 736000 }, { "epoch": 21.62, "grad_norm": 2.950507164001465, "learning_rate": 8.392186473420476e-05, "loss": 1.1825, "step": 736500 }, { "epoch": 21.64, "grad_norm": 2.338834047317505, "learning_rate": 8.37747727428734e-05, "loss": 1.1787, "step": 737000 }, { "epoch": 21.65, "grad_norm": 2.806655168533325, "learning_rate": 8.362768075154201e-05, "loss": 1.1988, "step": 737500 }, { "epoch": 21.67, "grad_norm": 2.509188652038574, "learning_rate": 8.348058876021062e-05, "loss": 1.1678, "step": 738000 }, { "epoch": 21.68, "grad_norm": 2.3886609077453613, "learning_rate": 8.333349676887924e-05, "loss": 1.2138, "step": 738500 }, { "epoch": 21.7, "grad_norm": 8.767037391662598, "learning_rate": 8.318640477754788e-05, "loss": 1.1682, "step": 739000 }, { "epoch": 21.71, "grad_norm": 2.7434234619140625, "learning_rate": 8.30393127862165e-05, "loss": 1.1675, "step": 739500 }, { "epoch": 21.73, "grad_norm": 3.837946891784668, "learning_rate": 8.28922207948851e-05, "loss": 1.1746, "step": 740000 }, { "epoch": 21.74, "grad_norm": 2.2261147499084473, "learning_rate": 8.274512880355373e-05, "loss": 1.1805, "step": 740500 }, { "epoch": 21.76, "grad_norm": 3.7343368530273438, "learning_rate": 8.259803681222236e-05, "loss": 1.1829, "step": 741000 }, { "epoch": 21.77, "grad_norm": 2.696150779724121, "learning_rate": 8.245094482089098e-05, "loss": 1.1791, "step": 741500 }, { "epoch": 21.79, "grad_norm": 3.374664783477783, "learning_rate": 8.230385282955959e-05, "loss": 1.186, "step": 742000 }, { "epoch": 21.8, "grad_norm": 2.0599725246429443, "learning_rate": 8.215676083822821e-05, "loss": 1.1704, "step": 742500 }, { "epoch": 21.81, "grad_norm": 2.960590362548828, "learning_rate": 8.200966884689685e-05, "loss": 1.2078, "step": 743000 }, { "epoch": 21.83, "grad_norm": 3.514355421066284, "learning_rate": 8.186257685556547e-05, "loss": 1.1853, "step": 743500 }, { "epoch": 21.84, "grad_norm": 3.7044436931610107, "learning_rate": 8.171548486423408e-05, "loss": 1.1775, "step": 744000 }, { "epoch": 21.86, "grad_norm": 2.7482707500457764, "learning_rate": 8.15683928729027e-05, "loss": 1.1927, "step": 744500 }, { "epoch": 21.87, "grad_norm": 3.462541103363037, "learning_rate": 8.142130088157133e-05, "loss": 1.155, "step": 745000 }, { "epoch": 21.89, "grad_norm": 1.3520176410675049, "learning_rate": 8.127420889023995e-05, "loss": 1.1887, "step": 745500 }, { "epoch": 21.9, "grad_norm": 3.2204971313476562, "learning_rate": 8.112711689890857e-05, "loss": 1.17, "step": 746000 }, { "epoch": 21.92, "grad_norm": 2.8493075370788574, "learning_rate": 8.098002490757718e-05, "loss": 1.1804, "step": 746500 }, { "epoch": 21.93, "grad_norm": 28.339750289916992, "learning_rate": 8.083293291624582e-05, "loss": 1.1832, "step": 747000 }, { "epoch": 21.95, "grad_norm": 3.3736674785614014, "learning_rate": 8.068584092491444e-05, "loss": 1.1903, "step": 747500 }, { "epoch": 21.96, "grad_norm": 2.166250228881836, "learning_rate": 8.053874893358306e-05, "loss": 1.1712, "step": 748000 }, { "epoch": 21.98, "grad_norm": 3.713568925857544, "learning_rate": 8.039165694225167e-05, "loss": 1.1896, "step": 748500 }, { "epoch": 21.99, "grad_norm": 3.077561140060425, "learning_rate": 8.02445649509203e-05, "loss": 1.1999, "step": 749000 }, { "epoch": 22.01, "grad_norm": 1.1793521642684937, "learning_rate": 8.009747295958892e-05, "loss": 1.166, "step": 749500 }, { "epoch": 22.02, "grad_norm": 1.4397152662277222, "learning_rate": 7.995038096825754e-05, "loss": 1.1327, "step": 750000 }, { "epoch": 22.03, "grad_norm": 6.889225959777832, "learning_rate": 7.980328897692615e-05, "loss": 1.1533, "step": 750500 }, { "epoch": 22.05, "grad_norm": 1.7650978565216064, "learning_rate": 7.965619698559479e-05, "loss": 1.1301, "step": 751000 }, { "epoch": 22.06, "grad_norm": 1.922248125076294, "learning_rate": 7.95091049942634e-05, "loss": 1.1313, "step": 751500 }, { "epoch": 22.08, "grad_norm": 9.232244491577148, "learning_rate": 7.936201300293203e-05, "loss": 1.1601, "step": 752000 }, { "epoch": 22.09, "grad_norm": 2.162410259246826, "learning_rate": 7.921492101160063e-05, "loss": 1.1394, "step": 752500 }, { "epoch": 22.11, "grad_norm": 2.0230438709259033, "learning_rate": 7.906782902026927e-05, "loss": 1.131, "step": 753000 }, { "epoch": 22.12, "grad_norm": 3.8333919048309326, "learning_rate": 7.892073702893789e-05, "loss": 1.1264, "step": 753500 }, { "epoch": 22.14, "grad_norm": 1.487121343612671, "learning_rate": 7.877364503760651e-05, "loss": 1.1127, "step": 754000 }, { "epoch": 22.15, "grad_norm": 2.5997982025146484, "learning_rate": 7.862655304627512e-05, "loss": 1.1072, "step": 754500 }, { "epoch": 22.17, "grad_norm": 2.5122156143188477, "learning_rate": 7.847946105494375e-05, "loss": 1.1091, "step": 755000 }, { "epoch": 22.18, "grad_norm": 2.6082444190979004, "learning_rate": 7.833236906361238e-05, "loss": 1.142, "step": 755500 }, { "epoch": 22.2, "grad_norm": 8.16508960723877, "learning_rate": 7.8185277072281e-05, "loss": 1.1247, "step": 756000 }, { "epoch": 22.21, "grad_norm": 3.061530113220215, "learning_rate": 7.803818508094962e-05, "loss": 1.1526, "step": 756500 }, { "epoch": 22.23, "grad_norm": 3.9026131629943848, "learning_rate": 7.789109308961824e-05, "loss": 1.152, "step": 757000 }, { "epoch": 22.24, "grad_norm": 2.3439905643463135, "learning_rate": 7.774400109828686e-05, "loss": 1.1416, "step": 757500 }, { "epoch": 22.26, "grad_norm": 0.9024741053581238, "learning_rate": 7.759690910695548e-05, "loss": 1.1272, "step": 758000 }, { "epoch": 22.27, "grad_norm": 3.3680572509765625, "learning_rate": 7.74498171156241e-05, "loss": 1.1445, "step": 758500 }, { "epoch": 22.28, "grad_norm": 1.695378065109253, "learning_rate": 7.730272512429274e-05, "loss": 1.1373, "step": 759000 }, { "epoch": 22.3, "grad_norm": 4.116308689117432, "learning_rate": 7.715563313296134e-05, "loss": 1.1453, "step": 759500 }, { "epoch": 22.31, "grad_norm": 2.014496088027954, "learning_rate": 7.700854114162997e-05, "loss": 1.146, "step": 760000 }, { "epoch": 22.33, "grad_norm": 1.5276461839675903, "learning_rate": 7.686144915029859e-05, "loss": 1.13, "step": 760500 }, { "epoch": 22.34, "grad_norm": 2.514599084854126, "learning_rate": 7.671435715896722e-05, "loss": 1.142, "step": 761000 }, { "epoch": 22.36, "grad_norm": 2.2747581005096436, "learning_rate": 7.656726516763583e-05, "loss": 1.1552, "step": 761500 }, { "epoch": 22.37, "grad_norm": 2.406036615371704, "learning_rate": 7.642017317630445e-05, "loss": 1.1453, "step": 762000 }, { "epoch": 22.39, "grad_norm": 2.231992721557617, "learning_rate": 7.627308118497307e-05, "loss": 1.1386, "step": 762500 }, { "epoch": 22.4, "grad_norm": 3.1717188358306885, "learning_rate": 7.61259891936417e-05, "loss": 1.1549, "step": 763000 }, { "epoch": 22.42, "grad_norm": 22.27747917175293, "learning_rate": 7.597889720231031e-05, "loss": 1.1142, "step": 763500 }, { "epoch": 22.43, "grad_norm": 2.7918033599853516, "learning_rate": 7.583180521097893e-05, "loss": 1.1108, "step": 764000 }, { "epoch": 22.45, "grad_norm": 1.4565542936325073, "learning_rate": 7.568471321964757e-05, "loss": 1.1509, "step": 764500 }, { "epoch": 22.46, "grad_norm": 3.1026453971862793, "learning_rate": 7.553762122831619e-05, "loss": 1.1513, "step": 765000 }, { "epoch": 22.48, "grad_norm": 1.6490851640701294, "learning_rate": 7.53905292369848e-05, "loss": 1.1242, "step": 765500 }, { "epoch": 22.49, "grad_norm": 1.7027279138565063, "learning_rate": 7.524343724565342e-05, "loss": 1.1561, "step": 766000 }, { "epoch": 22.5, "grad_norm": 1.2775441408157349, "learning_rate": 7.509634525432205e-05, "loss": 1.1168, "step": 766500 }, { "epoch": 22.52, "grad_norm": 5.64236307144165, "learning_rate": 7.494925326299067e-05, "loss": 1.1465, "step": 767000 }, { "epoch": 22.53, "grad_norm": 2.019430637359619, "learning_rate": 7.480216127165928e-05, "loss": 1.1162, "step": 767500 }, { "epoch": 22.55, "grad_norm": 1.3938747644424438, "learning_rate": 7.465506928032792e-05, "loss": 1.1253, "step": 768000 }, { "epoch": 22.56, "grad_norm": 2.122945785522461, "learning_rate": 7.450797728899654e-05, "loss": 1.1409, "step": 768500 }, { "epoch": 22.58, "grad_norm": 1.826390266418457, "learning_rate": 7.436088529766516e-05, "loss": 1.1473, "step": 769000 }, { "epoch": 22.59, "grad_norm": 1.8003076314926147, "learning_rate": 7.421379330633378e-05, "loss": 1.1827, "step": 769500 }, { "epoch": 22.61, "grad_norm": 2.6314899921417236, "learning_rate": 7.40667013150024e-05, "loss": 1.1426, "step": 770000 }, { "epoch": 22.62, "grad_norm": 2.7484140396118164, "learning_rate": 7.391960932367102e-05, "loss": 1.1603, "step": 770500 }, { "epoch": 22.64, "grad_norm": 1.5445499420166016, "learning_rate": 7.377251733233964e-05, "loss": 1.1609, "step": 771000 }, { "epoch": 22.65, "grad_norm": 2.3308262825012207, "learning_rate": 7.362542534100826e-05, "loss": 1.1126, "step": 771500 }, { "epoch": 22.67, "grad_norm": 2.675467014312744, "learning_rate": 7.347833334967689e-05, "loss": 1.1717, "step": 772000 }, { "epoch": 22.68, "grad_norm": 7.849360466003418, "learning_rate": 7.333124135834551e-05, "loss": 1.1304, "step": 772500 }, { "epoch": 22.7, "grad_norm": 1.596022129058838, "learning_rate": 7.318414936701413e-05, "loss": 1.1563, "step": 773000 }, { "epoch": 22.71, "grad_norm": 1.8678432703018188, "learning_rate": 7.303705737568275e-05, "loss": 1.1668, "step": 773500 }, { "epoch": 22.72, "grad_norm": 1.3757662773132324, "learning_rate": 7.288996538435137e-05, "loss": 1.1484, "step": 774000 }, { "epoch": 22.74, "grad_norm": 2.819344997406006, "learning_rate": 7.274287339301999e-05, "loss": 1.1213, "step": 774500 }, { "epoch": 22.75, "grad_norm": 2.370375871658325, "learning_rate": 7.259578140168861e-05, "loss": 1.1654, "step": 775000 }, { "epoch": 22.77, "grad_norm": 3.6313276290893555, "learning_rate": 7.244868941035723e-05, "loss": 1.13, "step": 775500 }, { "epoch": 22.78, "grad_norm": 1.737453579902649, "learning_rate": 7.230159741902585e-05, "loss": 1.1283, "step": 776000 }, { "epoch": 22.8, "grad_norm": 2.530315637588501, "learning_rate": 7.215450542769448e-05, "loss": 1.1413, "step": 776500 }, { "epoch": 22.81, "grad_norm": 2.049701452255249, "learning_rate": 7.20074134363631e-05, "loss": 1.1863, "step": 777000 }, { "epoch": 22.83, "grad_norm": 2.112698554992676, "learning_rate": 7.186032144503172e-05, "loss": 1.1217, "step": 777500 }, { "epoch": 22.84, "grad_norm": 1.982596755027771, "learning_rate": 7.171322945370034e-05, "loss": 1.151, "step": 778000 }, { "epoch": 22.86, "grad_norm": 2.8483023643493652, "learning_rate": 7.156613746236896e-05, "loss": 1.1371, "step": 778500 }, { "epoch": 22.87, "grad_norm": 2.3431105613708496, "learning_rate": 7.141904547103758e-05, "loss": 1.1549, "step": 779000 }, { "epoch": 22.89, "grad_norm": 2.256725549697876, "learning_rate": 7.12719534797062e-05, "loss": 1.1387, "step": 779500 }, { "epoch": 22.9, "grad_norm": 3.2769088745117188, "learning_rate": 7.112486148837482e-05, "loss": 1.1365, "step": 780000 }, { "epoch": 22.92, "grad_norm": 7.680245399475098, "learning_rate": 7.097776949704344e-05, "loss": 1.1524, "step": 780500 }, { "epoch": 22.93, "grad_norm": 2.0272703170776367, "learning_rate": 7.083067750571207e-05, "loss": 1.1483, "step": 781000 }, { "epoch": 22.95, "grad_norm": 3.6463356018066406, "learning_rate": 7.068358551438069e-05, "loss": 1.1188, "step": 781500 }, { "epoch": 22.96, "grad_norm": 2.499232769012451, "learning_rate": 7.053649352304931e-05, "loss": 1.1434, "step": 782000 }, { "epoch": 22.97, "grad_norm": 2.1540205478668213, "learning_rate": 7.038940153171793e-05, "loss": 1.1394, "step": 782500 }, { "epoch": 22.99, "grad_norm": 2.569671154022217, "learning_rate": 7.024230954038655e-05, "loss": 1.1581, "step": 783000 }, { "epoch": 23.0, "grad_norm": 5.3423333168029785, "learning_rate": 7.009521754905517e-05, "loss": 1.1392, "step": 783500 }, { "epoch": 23.02, "grad_norm": 1.8991897106170654, "learning_rate": 6.994812555772379e-05, "loss": 1.0877, "step": 784000 }, { "epoch": 23.03, "grad_norm": 2.094095230102539, "learning_rate": 6.980103356639241e-05, "loss": 1.1106, "step": 784500 }, { "epoch": 23.05, "grad_norm": 2.4271163940429688, "learning_rate": 6.965394157506104e-05, "loss": 1.0893, "step": 785000 }, { "epoch": 23.06, "grad_norm": 3.5016427040100098, "learning_rate": 6.950684958372966e-05, "loss": 1.1119, "step": 785500 }, { "epoch": 23.08, "grad_norm": 3.388533592224121, "learning_rate": 6.935975759239828e-05, "loss": 1.1135, "step": 786000 }, { "epoch": 23.09, "grad_norm": 1.610478401184082, "learning_rate": 6.92126656010669e-05, "loss": 1.1064, "step": 786500 }, { "epoch": 23.11, "grad_norm": 2.6857564449310303, "learning_rate": 6.906557360973552e-05, "loss": 1.105, "step": 787000 }, { "epoch": 23.12, "grad_norm": 3.785341739654541, "learning_rate": 6.891848161840414e-05, "loss": 1.1109, "step": 787500 }, { "epoch": 23.14, "grad_norm": 1.6604160070419312, "learning_rate": 6.877138962707276e-05, "loss": 1.1124, "step": 788000 }, { "epoch": 23.15, "grad_norm": 1.9755994081497192, "learning_rate": 6.862429763574138e-05, "loss": 1.1034, "step": 788500 }, { "epoch": 23.17, "grad_norm": 2.989527940750122, "learning_rate": 6.847720564441e-05, "loss": 1.1055, "step": 789000 }, { "epoch": 23.18, "grad_norm": 2.2718911170959473, "learning_rate": 6.833011365307863e-05, "loss": 1.104, "step": 789500 }, { "epoch": 23.19, "grad_norm": 4.303855895996094, "learning_rate": 6.818302166174725e-05, "loss": 1.1174, "step": 790000 }, { "epoch": 23.21, "grad_norm": 2.6747548580169678, "learning_rate": 6.803592967041587e-05, "loss": 1.1046, "step": 790500 }, { "epoch": 23.22, "grad_norm": 2.6597251892089844, "learning_rate": 6.78888376790845e-05, "loss": 1.1246, "step": 791000 }, { "epoch": 23.24, "grad_norm": 1.6096960306167603, "learning_rate": 6.774174568775311e-05, "loss": 1.1103, "step": 791500 }, { "epoch": 23.25, "grad_norm": 1.8474167585372925, "learning_rate": 6.759465369642174e-05, "loss": 1.1136, "step": 792000 }, { "epoch": 23.27, "grad_norm": 6.043608665466309, "learning_rate": 6.744756170509035e-05, "loss": 1.1118, "step": 792500 }, { "epoch": 23.28, "grad_norm": 10.027447700500488, "learning_rate": 6.730046971375899e-05, "loss": 1.115, "step": 793000 }, { "epoch": 23.3, "grad_norm": 2.5558454990386963, "learning_rate": 6.71533777224276e-05, "loss": 1.0993, "step": 793500 }, { "epoch": 23.31, "grad_norm": 2.075737476348877, "learning_rate": 6.700628573109623e-05, "loss": 1.1111, "step": 794000 }, { "epoch": 23.33, "grad_norm": 2.4479551315307617, "learning_rate": 6.685919373976484e-05, "loss": 1.0848, "step": 794500 }, { "epoch": 23.34, "grad_norm": 2.503129243850708, "learning_rate": 6.671210174843347e-05, "loss": 1.1315, "step": 795000 }, { "epoch": 23.36, "grad_norm": 2.2025949954986572, "learning_rate": 6.656500975710208e-05, "loss": 1.1061, "step": 795500 }, { "epoch": 23.37, "grad_norm": 2.5942320823669434, "learning_rate": 6.641791776577071e-05, "loss": 1.1244, "step": 796000 }, { "epoch": 23.39, "grad_norm": 2.2347915172576904, "learning_rate": 6.627082577443932e-05, "loss": 1.0868, "step": 796500 }, { "epoch": 23.4, "grad_norm": 2.3821372985839844, "learning_rate": 6.612373378310796e-05, "loss": 1.1175, "step": 797000 }, { "epoch": 23.41, "grad_norm": 2.4445548057556152, "learning_rate": 6.597664179177656e-05, "loss": 1.1163, "step": 797500 }, { "epoch": 23.43, "grad_norm": 4.338953018188477, "learning_rate": 6.58295498004452e-05, "loss": 1.104, "step": 798000 }, { "epoch": 23.44, "grad_norm": 1.766503930091858, "learning_rate": 6.56824578091138e-05, "loss": 1.1178, "step": 798500 }, { "epoch": 23.46, "grad_norm": 1.7907469272613525, "learning_rate": 6.553536581778244e-05, "loss": 1.1183, "step": 799000 }, { "epoch": 23.47, "grad_norm": 1.722770094871521, "learning_rate": 6.538827382645105e-05, "loss": 1.1009, "step": 799500 }, { "epoch": 23.49, "grad_norm": 1.7940298318862915, "learning_rate": 6.524118183511968e-05, "loss": 1.0938, "step": 800000 }, { "epoch": 23.5, "grad_norm": 1.5882244110107422, "learning_rate": 6.50940898437883e-05, "loss": 1.1024, "step": 800500 }, { "epoch": 23.52, "grad_norm": 1.7875205278396606, "learning_rate": 6.494699785245692e-05, "loss": 1.092, "step": 801000 }, { "epoch": 23.53, "grad_norm": 2.0258092880249023, "learning_rate": 6.479990586112555e-05, "loss": 1.1136, "step": 801500 }, { "epoch": 23.55, "grad_norm": 1.6038285493850708, "learning_rate": 6.465281386979417e-05, "loss": 1.1013, "step": 802000 }, { "epoch": 23.56, "grad_norm": 2.411367654800415, "learning_rate": 6.450572187846279e-05, "loss": 1.0971, "step": 802500 }, { "epoch": 23.58, "grad_norm": 18.325363159179688, "learning_rate": 6.435862988713141e-05, "loss": 1.1292, "step": 803000 }, { "epoch": 23.59, "grad_norm": 2.520692825317383, "learning_rate": 6.421153789580003e-05, "loss": 1.1108, "step": 803500 }, { "epoch": 23.61, "grad_norm": 2.115523099899292, "learning_rate": 6.406444590446865e-05, "loss": 1.0878, "step": 804000 }, { "epoch": 23.62, "grad_norm": 2.4668285846710205, "learning_rate": 6.391735391313727e-05, "loss": 1.1089, "step": 804500 }, { "epoch": 23.64, "grad_norm": 2.2170443534851074, "learning_rate": 6.37702619218059e-05, "loss": 1.1189, "step": 805000 }, { "epoch": 23.65, "grad_norm": 2.2153406143188477, "learning_rate": 6.362316993047451e-05, "loss": 1.1053, "step": 805500 }, { "epoch": 23.66, "grad_norm": 2.937535285949707, "learning_rate": 6.347607793914314e-05, "loss": 1.108, "step": 806000 }, { "epoch": 23.68, "grad_norm": 1.9660471677780151, "learning_rate": 6.332898594781176e-05, "loss": 1.1203, "step": 806500 }, { "epoch": 23.69, "grad_norm": 3.5854222774505615, "learning_rate": 6.318189395648038e-05, "loss": 1.1082, "step": 807000 }, { "epoch": 23.71, "grad_norm": 6.527500152587891, "learning_rate": 6.3034801965149e-05, "loss": 1.1146, "step": 807500 }, { "epoch": 23.72, "grad_norm": 7.003795623779297, "learning_rate": 6.288770997381762e-05, "loss": 1.1137, "step": 808000 }, { "epoch": 23.74, "grad_norm": 2.606776714324951, "learning_rate": 6.274061798248624e-05, "loss": 1.0925, "step": 808500 }, { "epoch": 23.75, "grad_norm": 3.0098581314086914, "learning_rate": 6.259352599115486e-05, "loss": 1.1041, "step": 809000 }, { "epoch": 23.77, "grad_norm": 1.7633291482925415, "learning_rate": 6.244643399982348e-05, "loss": 1.1257, "step": 809500 }, { "epoch": 23.78, "grad_norm": 2.6187331676483154, "learning_rate": 6.22993420084921e-05, "loss": 1.0956, "step": 810000 }, { "epoch": 23.8, "grad_norm": 1.919297456741333, "learning_rate": 6.215225001716073e-05, "loss": 1.0954, "step": 810500 }, { "epoch": 23.81, "grad_norm": 3.422133445739746, "learning_rate": 6.200515802582935e-05, "loss": 1.0997, "step": 811000 }, { "epoch": 23.83, "grad_norm": 1.8277616500854492, "learning_rate": 6.185806603449797e-05, "loss": 1.1357, "step": 811500 }, { "epoch": 23.84, "grad_norm": 1.641455054283142, "learning_rate": 6.171097404316659e-05, "loss": 1.1186, "step": 812000 }, { "epoch": 23.86, "grad_norm": 2.629831075668335, "learning_rate": 6.156388205183521e-05, "loss": 1.0844, "step": 812500 }, { "epoch": 23.87, "grad_norm": 2.9981462955474854, "learning_rate": 6.141679006050383e-05, "loss": 1.1323, "step": 813000 }, { "epoch": 23.88, "grad_norm": 7.224218845367432, "learning_rate": 6.126969806917245e-05, "loss": 1.1091, "step": 813500 }, { "epoch": 23.9, "grad_norm": 1.445144772529602, "learning_rate": 6.112260607784107e-05, "loss": 1.0885, "step": 814000 }, { "epoch": 23.91, "grad_norm": 6.933565616607666, "learning_rate": 6.0975514086509695e-05, "loss": 1.1264, "step": 814500 }, { "epoch": 23.93, "grad_norm": 1.5660400390625, "learning_rate": 6.082842209517832e-05, "loss": 1.102, "step": 815000 }, { "epoch": 23.94, "grad_norm": 3.0154061317443848, "learning_rate": 6.068133010384694e-05, "loss": 1.1043, "step": 815500 }, { "epoch": 23.96, "grad_norm": 4.825801372528076, "learning_rate": 6.0534238112515565e-05, "loss": 1.1289, "step": 816000 }, { "epoch": 23.97, "grad_norm": 2.10929012298584, "learning_rate": 6.038714612118418e-05, "loss": 1.112, "step": 816500 }, { "epoch": 23.99, "grad_norm": 15.725834846496582, "learning_rate": 6.024005412985281e-05, "loss": 1.0857, "step": 817000 }, { "epoch": 24.0, "grad_norm": 3.745626926422119, "learning_rate": 6.009296213852142e-05, "loss": 1.0957, "step": 817500 }, { "epoch": 24.02, "grad_norm": 1.5650004148483276, "learning_rate": 5.994587014719005e-05, "loss": 1.0553, "step": 818000 }, { "epoch": 24.03, "grad_norm": 1.279994249343872, "learning_rate": 5.9798778155858664e-05, "loss": 1.0524, "step": 818500 }, { "epoch": 24.05, "grad_norm": 1.6556503772735596, "learning_rate": 5.965168616452729e-05, "loss": 1.0709, "step": 819000 }, { "epoch": 24.06, "grad_norm": 1.5529954433441162, "learning_rate": 5.9504594173195906e-05, "loss": 1.0952, "step": 819500 }, { "epoch": 24.08, "grad_norm": 1.7145531177520752, "learning_rate": 5.9357502181864534e-05, "loss": 1.0526, "step": 820000 }, { "epoch": 24.09, "grad_norm": 1.5692338943481445, "learning_rate": 5.921041019053315e-05, "loss": 1.0493, "step": 820500 }, { "epoch": 24.1, "grad_norm": 37.72039031982422, "learning_rate": 5.9063318199201776e-05, "loss": 1.068, "step": 821000 }, { "epoch": 24.12, "grad_norm": 2.0122039318084717, "learning_rate": 5.891622620787039e-05, "loss": 1.0775, "step": 821500 }, { "epoch": 24.13, "grad_norm": 5.577430248260498, "learning_rate": 5.876913421653902e-05, "loss": 1.1098, "step": 822000 }, { "epoch": 24.15, "grad_norm": 3.4199016094207764, "learning_rate": 5.8622042225207647e-05, "loss": 1.0893, "step": 822500 }, { "epoch": 24.16, "grad_norm": 3.654630184173584, "learning_rate": 5.847495023387626e-05, "loss": 1.0719, "step": 823000 }, { "epoch": 24.18, "grad_norm": 2.0985825061798096, "learning_rate": 5.832785824254489e-05, "loss": 1.0561, "step": 823500 }, { "epoch": 24.19, "grad_norm": 3.724655866622925, "learning_rate": 5.81807662512135e-05, "loss": 1.047, "step": 824000 }, { "epoch": 24.21, "grad_norm": 2.0398764610290527, "learning_rate": 5.803367425988213e-05, "loss": 1.0806, "step": 824500 }, { "epoch": 24.22, "grad_norm": 2.9711148738861084, "learning_rate": 5.7886582268550745e-05, "loss": 1.104, "step": 825000 }, { "epoch": 24.24, "grad_norm": 1.4870119094848633, "learning_rate": 5.773949027721937e-05, "loss": 1.0812, "step": 825500 }, { "epoch": 24.25, "grad_norm": 2.184882640838623, "learning_rate": 5.759239828588799e-05, "loss": 1.071, "step": 826000 }, { "epoch": 24.27, "grad_norm": 2.135782241821289, "learning_rate": 5.7445306294556616e-05, "loss": 1.081, "step": 826500 }, { "epoch": 24.28, "grad_norm": 1.952862024307251, "learning_rate": 5.729821430322523e-05, "loss": 1.0688, "step": 827000 }, { "epoch": 24.3, "grad_norm": 3.28511381149292, "learning_rate": 5.715112231189386e-05, "loss": 1.0719, "step": 827500 }, { "epoch": 24.31, "grad_norm": 3.095506191253662, "learning_rate": 5.700403032056247e-05, "loss": 1.1089, "step": 828000 }, { "epoch": 24.33, "grad_norm": 1.744498372077942, "learning_rate": 5.68569383292311e-05, "loss": 1.0666, "step": 828500 }, { "epoch": 24.34, "grad_norm": 1.5370303392410278, "learning_rate": 5.6709846337899714e-05, "loss": 1.0834, "step": 829000 }, { "epoch": 24.35, "grad_norm": 3.65459942817688, "learning_rate": 5.656275434656834e-05, "loss": 1.0767, "step": 829500 }, { "epoch": 24.37, "grad_norm": 2.0333411693573, "learning_rate": 5.641566235523696e-05, "loss": 1.0965, "step": 830000 }, { "epoch": 24.38, "grad_norm": 2.9158904552459717, "learning_rate": 5.6268570363905585e-05, "loss": 1.0726, "step": 830500 }, { "epoch": 24.4, "grad_norm": 2.0873067378997803, "learning_rate": 5.6121478372574206e-05, "loss": 1.0664, "step": 831000 }, { "epoch": 24.41, "grad_norm": 2.0477983951568604, "learning_rate": 5.597438638124283e-05, "loss": 1.0588, "step": 831500 }, { "epoch": 24.43, "grad_norm": 1.9659167528152466, "learning_rate": 5.582729438991145e-05, "loss": 1.0986, "step": 832000 }, { "epoch": 24.44, "grad_norm": 2.130729913711548, "learning_rate": 5.568020239858007e-05, "loss": 1.0713, "step": 832500 }, { "epoch": 24.46, "grad_norm": 99.76852416992188, "learning_rate": 5.553311040724869e-05, "loss": 1.0803, "step": 833000 }, { "epoch": 24.47, "grad_norm": 5.594266414642334, "learning_rate": 5.538601841591731e-05, "loss": 1.08, "step": 833500 }, { "epoch": 24.49, "grad_norm": 2.3737406730651855, "learning_rate": 5.523892642458593e-05, "loss": 1.0931, "step": 834000 }, { "epoch": 24.5, "grad_norm": 1.6415624618530273, "learning_rate": 5.5091834433254554e-05, "loss": 1.0798, "step": 834500 }, { "epoch": 24.52, "grad_norm": 7.322782516479492, "learning_rate": 5.4944742441923175e-05, "loss": 1.0865, "step": 835000 }, { "epoch": 24.53, "grad_norm": 2.1509947776794434, "learning_rate": 5.4797650450591796e-05, "loss": 1.0782, "step": 835500 }, { "epoch": 24.55, "grad_norm": 2.485328197479248, "learning_rate": 5.465055845926042e-05, "loss": 1.0585, "step": 836000 }, { "epoch": 24.56, "grad_norm": 2.5374958515167236, "learning_rate": 5.450346646792904e-05, "loss": 1.0577, "step": 836500 }, { "epoch": 24.57, "grad_norm": 2.196706533432007, "learning_rate": 5.435637447659766e-05, "loss": 1.0894, "step": 837000 }, { "epoch": 24.59, "grad_norm": 2.9057724475860596, "learning_rate": 5.420928248526629e-05, "loss": 1.075, "step": 837500 }, { "epoch": 24.6, "grad_norm": 2.4338793754577637, "learning_rate": 5.40621904939349e-05, "loss": 1.0829, "step": 838000 }, { "epoch": 24.62, "grad_norm": 2.225640058517456, "learning_rate": 5.391509850260353e-05, "loss": 1.0798, "step": 838500 }, { "epoch": 24.63, "grad_norm": 2.3546741008758545, "learning_rate": 5.3768006511272144e-05, "loss": 1.0887, "step": 839000 }, { "epoch": 24.65, "grad_norm": 1.3033528327941895, "learning_rate": 5.362091451994077e-05, "loss": 1.0809, "step": 839500 }, { "epoch": 24.66, "grad_norm": 2.5136497020721436, "learning_rate": 5.3473822528609386e-05, "loss": 1.0851, "step": 840000 }, { "epoch": 24.68, "grad_norm": 2.0871777534484863, "learning_rate": 5.3326730537278014e-05, "loss": 1.0744, "step": 840500 }, { "epoch": 24.69, "grad_norm": 5.352667808532715, "learning_rate": 5.317963854594663e-05, "loss": 1.102, "step": 841000 }, { "epoch": 24.71, "grad_norm": 1.8060741424560547, "learning_rate": 5.3032546554615256e-05, "loss": 1.0541, "step": 841500 }, { "epoch": 24.72, "grad_norm": 2.5160653591156006, "learning_rate": 5.288545456328387e-05, "loss": 1.077, "step": 842000 }, { "epoch": 24.74, "grad_norm": 4.854517936706543, "learning_rate": 5.27383625719525e-05, "loss": 1.084, "step": 842500 }, { "epoch": 24.75, "grad_norm": 7.554323196411133, "learning_rate": 5.259127058062111e-05, "loss": 1.0974, "step": 843000 }, { "epoch": 24.77, "grad_norm": 1.8624228239059448, "learning_rate": 5.244417858928974e-05, "loss": 1.0793, "step": 843500 }, { "epoch": 24.78, "grad_norm": 1.6777387857437134, "learning_rate": 5.2297086597958355e-05, "loss": 1.0746, "step": 844000 }, { "epoch": 24.79, "grad_norm": 1.9839929342269897, "learning_rate": 5.214999460662698e-05, "loss": 1.0813, "step": 844500 }, { "epoch": 24.81, "grad_norm": 1.5621155500411987, "learning_rate": 5.20029026152956e-05, "loss": 1.0936, "step": 845000 }, { "epoch": 24.82, "grad_norm": 2.5677313804626465, "learning_rate": 5.1855810623964225e-05, "loss": 1.0845, "step": 845500 }, { "epoch": 24.84, "grad_norm": 1.3594130277633667, "learning_rate": 5.170871863263284e-05, "loss": 1.0563, "step": 846000 }, { "epoch": 24.85, "grad_norm": 1.3905824422836304, "learning_rate": 5.156162664130147e-05, "loss": 1.0661, "step": 846500 }, { "epoch": 24.87, "grad_norm": 1.6563278436660767, "learning_rate": 5.141453464997008e-05, "loss": 1.079, "step": 847000 }, { "epoch": 24.88, "grad_norm": 2.1937708854675293, "learning_rate": 5.126744265863871e-05, "loss": 1.0714, "step": 847500 }, { "epoch": 24.9, "grad_norm": 1.874009609222412, "learning_rate": 5.112035066730733e-05, "loss": 1.064, "step": 848000 }, { "epoch": 24.91, "grad_norm": 1.4228936433792114, "learning_rate": 5.097325867597595e-05, "loss": 1.0954, "step": 848500 }, { "epoch": 24.93, "grad_norm": 1.7444462776184082, "learning_rate": 5.082616668464457e-05, "loss": 1.0921, "step": 849000 }, { "epoch": 24.94, "grad_norm": 186.56033325195312, "learning_rate": 5.0679074693313194e-05, "loss": 1.0757, "step": 849500 }, { "epoch": 24.96, "grad_norm": 4.014658451080322, "learning_rate": 5.0531982701981815e-05, "loss": 1.0864, "step": 850000 }, { "epoch": 24.97, "grad_norm": 1.8258497714996338, "learning_rate": 5.0384890710650436e-05, "loss": 1.0691, "step": 850500 }, { "epoch": 24.99, "grad_norm": 2.100780487060547, "learning_rate": 5.023779871931906e-05, "loss": 1.0722, "step": 851000 }, { "epoch": 25.0, "grad_norm": 2.250474452972412, "learning_rate": 5.009070672798768e-05, "loss": 1.0853, "step": 851500 }, { "epoch": 25.02, "grad_norm": 2.6746740341186523, "learning_rate": 4.99436147366563e-05, "loss": 1.0277, "step": 852000 }, { "epoch": 25.03, "grad_norm": 2.413872480392456, "learning_rate": 4.979652274532492e-05, "loss": 1.0804, "step": 852500 }, { "epoch": 25.04, "grad_norm": 3.563396453857422, "learning_rate": 4.964943075399354e-05, "loss": 1.0638, "step": 853000 }, { "epoch": 25.06, "grad_norm": 1.7918055057525635, "learning_rate": 4.950233876266217e-05, "loss": 1.0541, "step": 853500 }, { "epoch": 25.07, "grad_norm": 2.508848190307617, "learning_rate": 4.9355246771330784e-05, "loss": 1.0083, "step": 854000 }, { "epoch": 25.09, "grad_norm": 9.781774520874023, "learning_rate": 4.920815477999941e-05, "loss": 1.0236, "step": 854500 }, { "epoch": 25.1, "grad_norm": 2.6363613605499268, "learning_rate": 4.9061062788668026e-05, "loss": 1.0356, "step": 855000 }, { "epoch": 25.12, "grad_norm": 5.132301330566406, "learning_rate": 4.8913970797336654e-05, "loss": 1.0509, "step": 855500 }, { "epoch": 25.13, "grad_norm": 2.983057737350464, "learning_rate": 4.876687880600527e-05, "loss": 1.0486, "step": 856000 }, { "epoch": 25.15, "grad_norm": 3.175165891647339, "learning_rate": 4.8619786814673897e-05, "loss": 1.0521, "step": 856500 }, { "epoch": 25.16, "grad_norm": 5.045122146606445, "learning_rate": 4.847269482334251e-05, "loss": 1.0494, "step": 857000 }, { "epoch": 25.18, "grad_norm": 2.601104259490967, "learning_rate": 4.832560283201114e-05, "loss": 1.0301, "step": 857500 }, { "epoch": 25.19, "grad_norm": 2.546464443206787, "learning_rate": 4.817851084067975e-05, "loss": 1.0237, "step": 858000 }, { "epoch": 25.21, "grad_norm": 2.9643750190734863, "learning_rate": 4.803141884934838e-05, "loss": 1.0573, "step": 858500 }, { "epoch": 25.22, "grad_norm": 5.178891658782959, "learning_rate": 4.7884326858016995e-05, "loss": 1.0588, "step": 859000 }, { "epoch": 25.24, "grad_norm": 2.786492347717285, "learning_rate": 4.773723486668562e-05, "loss": 1.0537, "step": 859500 }, { "epoch": 25.25, "grad_norm": 3.3391730785369873, "learning_rate": 4.759014287535424e-05, "loss": 1.0452, "step": 860000 }, { "epoch": 25.26, "grad_norm": 2.8821425437927246, "learning_rate": 4.7443050884022866e-05, "loss": 1.0538, "step": 860500 }, { "epoch": 25.28, "grad_norm": 8.122342109680176, "learning_rate": 4.729595889269148e-05, "loss": 1.0326, "step": 861000 }, { "epoch": 25.29, "grad_norm": 5.042423725128174, "learning_rate": 4.714886690136011e-05, "loss": 1.0586, "step": 861500 }, { "epoch": 25.31, "grad_norm": 3.068471670150757, "learning_rate": 4.700177491002872e-05, "loss": 1.032, "step": 862000 }, { "epoch": 25.32, "grad_norm": 3.833307981491089, "learning_rate": 4.685468291869735e-05, "loss": 1.0581, "step": 862500 }, { "epoch": 25.34, "grad_norm": 2.662344455718994, "learning_rate": 4.6707590927365964e-05, "loss": 1.0526, "step": 863000 }, { "epoch": 25.35, "grad_norm": 3.626770257949829, "learning_rate": 4.656049893603459e-05, "loss": 1.0484, "step": 863500 }, { "epoch": 25.37, "grad_norm": 2.1823232173919678, "learning_rate": 4.6413406944703213e-05, "loss": 1.0524, "step": 864000 }, { "epoch": 25.38, "grad_norm": 2.051774740219116, "learning_rate": 4.6266314953371835e-05, "loss": 1.0406, "step": 864500 }, { "epoch": 25.4, "grad_norm": 9.218499183654785, "learning_rate": 4.6119222962040456e-05, "loss": 1.0605, "step": 865000 }, { "epoch": 25.41, "grad_norm": 2.4243879318237305, "learning_rate": 4.597213097070908e-05, "loss": 1.033, "step": 865500 }, { "epoch": 25.43, "grad_norm": 5.0782599449157715, "learning_rate": 4.58250389793777e-05, "loss": 1.059, "step": 866000 }, { "epoch": 25.44, "grad_norm": 2.4752907752990723, "learning_rate": 4.567794698804632e-05, "loss": 1.0348, "step": 866500 }, { "epoch": 25.46, "grad_norm": 3.958376407623291, "learning_rate": 4.553085499671494e-05, "loss": 1.0677, "step": 867000 }, { "epoch": 25.47, "grad_norm": 2.9827880859375, "learning_rate": 4.538376300538356e-05, "loss": 1.0371, "step": 867500 }, { "epoch": 25.48, "grad_norm": 4.071894645690918, "learning_rate": 4.523667101405218e-05, "loss": 1.0578, "step": 868000 }, { "epoch": 25.5, "grad_norm": 34.7899169921875, "learning_rate": 4.5089579022720804e-05, "loss": 1.0903, "step": 868500 }, { "epoch": 25.51, "grad_norm": 2.133716344833374, "learning_rate": 4.4942487031389425e-05, "loss": 1.0471, "step": 869000 }, { "epoch": 25.53, "grad_norm": 2.6652045249938965, "learning_rate": 4.4795395040058046e-05, "loss": 1.0358, "step": 869500 }, { "epoch": 25.54, "grad_norm": 2.98919677734375, "learning_rate": 4.464830304872667e-05, "loss": 1.0621, "step": 870000 }, { "epoch": 25.56, "grad_norm": 2.7026760578155518, "learning_rate": 4.4501211057395295e-05, "loss": 1.0621, "step": 870500 }, { "epoch": 25.57, "grad_norm": 3.262974977493286, "learning_rate": 4.435411906606391e-05, "loss": 1.0321, "step": 871000 }, { "epoch": 25.59, "grad_norm": 1.9586899280548096, "learning_rate": 4.420702707473254e-05, "loss": 1.0347, "step": 871500 }, { "epoch": 25.6, "grad_norm": 7.560853481292725, "learning_rate": 4.405993508340115e-05, "loss": 1.0221, "step": 872000 }, { "epoch": 25.62, "grad_norm": 3.056516408920288, "learning_rate": 4.391284309206978e-05, "loss": 1.043, "step": 872500 }, { "epoch": 25.63, "grad_norm": 2.7999415397644043, "learning_rate": 4.3765751100738394e-05, "loss": 1.0515, "step": 873000 }, { "epoch": 25.65, "grad_norm": 2.5458314418792725, "learning_rate": 4.361865910940702e-05, "loss": 1.0599, "step": 873500 }, { "epoch": 25.66, "grad_norm": 3.634589910507202, "learning_rate": 4.3471567118075636e-05, "loss": 1.0428, "step": 874000 }, { "epoch": 25.68, "grad_norm": 2.8104031085968018, "learning_rate": 4.3324475126744264e-05, "loss": 1.0396, "step": 874500 }, { "epoch": 25.69, "grad_norm": 2.319333791732788, "learning_rate": 4.317738313541288e-05, "loss": 1.0426, "step": 875000 }, { "epoch": 25.71, "grad_norm": 8.262571334838867, "learning_rate": 4.3030291144081506e-05, "loss": 1.0616, "step": 875500 }, { "epoch": 25.72, "grad_norm": 3.307812213897705, "learning_rate": 4.288319915275012e-05, "loss": 1.0706, "step": 876000 }, { "epoch": 25.73, "grad_norm": 7.4953508377075195, "learning_rate": 4.273610716141875e-05, "loss": 1.0457, "step": 876500 }, { "epoch": 25.75, "grad_norm": 3.5895800590515137, "learning_rate": 4.258901517008736e-05, "loss": 1.0219, "step": 877000 }, { "epoch": 25.76, "grad_norm": 3.023827075958252, "learning_rate": 4.244192317875599e-05, "loss": 1.05, "step": 877500 }, { "epoch": 25.78, "grad_norm": 2.1299614906311035, "learning_rate": 4.2294831187424605e-05, "loss": 1.0505, "step": 878000 }, { "epoch": 25.79, "grad_norm": 3.042360305786133, "learning_rate": 4.214773919609323e-05, "loss": 1.0481, "step": 878500 }, { "epoch": 25.81, "grad_norm": 2.3197755813598633, "learning_rate": 4.200064720476185e-05, "loss": 1.0181, "step": 879000 }, { "epoch": 25.82, "grad_norm": 3.230426788330078, "learning_rate": 4.1853555213430475e-05, "loss": 1.0327, "step": 879500 }, { "epoch": 25.84, "grad_norm": 7.685334205627441, "learning_rate": 4.1706463222099096e-05, "loss": 1.0334, "step": 880000 }, { "epoch": 25.85, "grad_norm": 3.2607569694519043, "learning_rate": 4.155937123076772e-05, "loss": 1.0342, "step": 880500 }, { "epoch": 25.87, "grad_norm": 3.470470428466797, "learning_rate": 4.1412279239436345e-05, "loss": 1.0508, "step": 881000 }, { "epoch": 25.88, "grad_norm": 16.141782760620117, "learning_rate": 4.126518724810496e-05, "loss": 1.046, "step": 881500 }, { "epoch": 25.9, "grad_norm": 2.900520086288452, "learning_rate": 4.111809525677359e-05, "loss": 1.0791, "step": 882000 }, { "epoch": 25.91, "grad_norm": 2.5581626892089844, "learning_rate": 4.09710032654422e-05, "loss": 1.0386, "step": 882500 }, { "epoch": 25.93, "grad_norm": 2.301318883895874, "learning_rate": 4.082391127411083e-05, "loss": 1.078, "step": 883000 }, { "epoch": 25.94, "grad_norm": 2.5168445110321045, "learning_rate": 4.0676819282779444e-05, "loss": 1.0524, "step": 883500 }, { "epoch": 25.95, "grad_norm": 3.2487659454345703, "learning_rate": 4.052972729144807e-05, "loss": 1.0476, "step": 884000 }, { "epoch": 25.97, "grad_norm": 3.8711748123168945, "learning_rate": 4.0382635300116686e-05, "loss": 1.0349, "step": 884500 }, { "epoch": 25.98, "grad_norm": 2.8647818565368652, "learning_rate": 4.0235543308785314e-05, "loss": 1.0681, "step": 885000 }, { "epoch": 26.0, "grad_norm": 2.307555675506592, "learning_rate": 4.008845131745393e-05, "loss": 1.0299, "step": 885500 }, { "epoch": 26.01, "grad_norm": 5.377994537353516, "learning_rate": 3.9941359326122557e-05, "loss": 1.0274, "step": 886000 }, { "epoch": 26.03, "grad_norm": 3.1154091358184814, "learning_rate": 3.979426733479118e-05, "loss": 1.0313, "step": 886500 }, { "epoch": 26.04, "grad_norm": 1.3295516967773438, "learning_rate": 3.96471753434598e-05, "loss": 0.9962, "step": 887000 }, { "epoch": 26.06, "grad_norm": 2.4110894203186035, "learning_rate": 3.950008335212842e-05, "loss": 1.0051, "step": 887500 }, { "epoch": 26.07, "grad_norm": 1.6697067022323608, "learning_rate": 3.935299136079704e-05, "loss": 1.0119, "step": 888000 }, { "epoch": 26.09, "grad_norm": 2.4017999172210693, "learning_rate": 3.920589936946566e-05, "loss": 1.0253, "step": 888500 }, { "epoch": 26.1, "grad_norm": 3.1681103706359863, "learning_rate": 3.905880737813428e-05, "loss": 1.0213, "step": 889000 }, { "epoch": 26.12, "grad_norm": 2.9184956550598145, "learning_rate": 3.8911715386802904e-05, "loss": 1.0298, "step": 889500 }, { "epoch": 26.13, "grad_norm": 1.6290541887283325, "learning_rate": 3.8764623395471526e-05, "loss": 1.0094, "step": 890000 }, { "epoch": 26.15, "grad_norm": 2.779973030090332, "learning_rate": 3.861753140414015e-05, "loss": 1.016, "step": 890500 }, { "epoch": 26.16, "grad_norm": 4.393610000610352, "learning_rate": 3.847043941280877e-05, "loss": 1.0131, "step": 891000 }, { "epoch": 26.17, "grad_norm": 2.503702402114868, "learning_rate": 3.832334742147739e-05, "loss": 1.0207, "step": 891500 }, { "epoch": 26.19, "grad_norm": 1.3769996166229248, "learning_rate": 3.817625543014601e-05, "loss": 1.0305, "step": 892000 }, { "epoch": 26.2, "grad_norm": 9.221471786499023, "learning_rate": 3.802916343881463e-05, "loss": 1.0258, "step": 892500 }, { "epoch": 26.22, "grad_norm": 1.706425666809082, "learning_rate": 3.788207144748326e-05, "loss": 0.9808, "step": 893000 }, { "epoch": 26.23, "grad_norm": 15.878872871398926, "learning_rate": 3.7734979456151873e-05, "loss": 1.0225, "step": 893500 }, { "epoch": 26.25, "grad_norm": 9.141250610351562, "learning_rate": 3.75878874648205e-05, "loss": 1.0234, "step": 894000 }, { "epoch": 26.26, "grad_norm": 1.698013424873352, "learning_rate": 3.7440795473489116e-05, "loss": 1.0207, "step": 894500 }, { "epoch": 26.28, "grad_norm": 2.8760321140289307, "learning_rate": 3.729370348215774e-05, "loss": 1.0054, "step": 895000 }, { "epoch": 26.29, "grad_norm": 10.192665100097656, "learning_rate": 3.714661149082636e-05, "loss": 1.0202, "step": 895500 }, { "epoch": 26.31, "grad_norm": 1.8373937606811523, "learning_rate": 3.699951949949498e-05, "loss": 1.0399, "step": 896000 }, { "epoch": 26.32, "grad_norm": 0.8489872217178345, "learning_rate": 3.68524275081636e-05, "loss": 1.0525, "step": 896500 }, { "epoch": 26.34, "grad_norm": 1.7799832820892334, "learning_rate": 3.670533551683222e-05, "loss": 1.0271, "step": 897000 }, { "epoch": 26.35, "grad_norm": 1.6424106359481812, "learning_rate": 3.655824352550084e-05, "loss": 1.0093, "step": 897500 }, { "epoch": 26.37, "grad_norm": 4.4057135581970215, "learning_rate": 3.6411151534169464e-05, "loss": 1.0214, "step": 898000 }, { "epoch": 26.38, "grad_norm": 1.971706509590149, "learning_rate": 3.6264059542838085e-05, "loss": 1.0316, "step": 898500 }, { "epoch": 26.39, "grad_norm": 1.934280276298523, "learning_rate": 3.6116967551506706e-05, "loss": 1.0368, "step": 899000 }, { "epoch": 26.41, "grad_norm": 1.6416791677474976, "learning_rate": 3.596987556017533e-05, "loss": 1.0123, "step": 899500 }, { "epoch": 26.42, "grad_norm": 1.6787991523742676, "learning_rate": 3.582278356884395e-05, "loss": 1.0026, "step": 900000 }, { "epoch": 26.44, "grad_norm": 2.6135306358337402, "learning_rate": 3.567569157751257e-05, "loss": 0.9887, "step": 900500 }, { "epoch": 26.45, "grad_norm": 1.4282991886138916, "learning_rate": 3.552859958618119e-05, "loss": 1.0442, "step": 901000 }, { "epoch": 26.47, "grad_norm": 1.8328664302825928, "learning_rate": 3.538150759484981e-05, "loss": 0.9983, "step": 901500 }, { "epoch": 26.48, "grad_norm": 2.659299850463867, "learning_rate": 3.523441560351843e-05, "loss": 1.0397, "step": 902000 }, { "epoch": 26.5, "grad_norm": 2.29178524017334, "learning_rate": 3.508732361218706e-05, "loss": 1.0182, "step": 902500 }, { "epoch": 26.51, "grad_norm": 9.058496475219727, "learning_rate": 3.494023162085568e-05, "loss": 1.0113, "step": 903000 }, { "epoch": 26.53, "grad_norm": 1.5530019998550415, "learning_rate": 3.47931396295243e-05, "loss": 1.0243, "step": 903500 }, { "epoch": 26.54, "grad_norm": 1.9570651054382324, "learning_rate": 3.4646047638192924e-05, "loss": 1.0012, "step": 904000 }, { "epoch": 26.56, "grad_norm": 1.154820203781128, "learning_rate": 3.4498955646861545e-05, "loss": 1.0231, "step": 904500 }, { "epoch": 26.57, "grad_norm": 1.5875574350357056, "learning_rate": 3.4351863655530166e-05, "loss": 1.0239, "step": 905000 }, { "epoch": 26.59, "grad_norm": 2.007080554962158, "learning_rate": 3.420477166419879e-05, "loss": 1.0119, "step": 905500 }, { "epoch": 26.6, "grad_norm": 1.6997588872909546, "learning_rate": 3.405767967286741e-05, "loss": 1.0235, "step": 906000 }, { "epoch": 26.62, "grad_norm": 2.8754279613494873, "learning_rate": 3.391058768153603e-05, "loss": 1.0152, "step": 906500 }, { "epoch": 26.63, "grad_norm": 3.1198441982269287, "learning_rate": 3.376349569020465e-05, "loss": 1.0351, "step": 907000 }, { "epoch": 26.64, "grad_norm": 11.747318267822266, "learning_rate": 3.361640369887327e-05, "loss": 1.0166, "step": 907500 }, { "epoch": 26.66, "grad_norm": 1.7430450916290283, "learning_rate": 3.346931170754189e-05, "loss": 1.0268, "step": 908000 }, { "epoch": 26.67, "grad_norm": 1.4505376815795898, "learning_rate": 3.3322219716210514e-05, "loss": 1.0297, "step": 908500 }, { "epoch": 26.69, "grad_norm": 4.441274642944336, "learning_rate": 3.3175127724879135e-05, "loss": 1.0145, "step": 909000 }, { "epoch": 26.7, "grad_norm": 4.974277973175049, "learning_rate": 3.302803573354776e-05, "loss": 1.0163, "step": 909500 }, { "epoch": 26.72, "grad_norm": 2.1050703525543213, "learning_rate": 3.2880943742216384e-05, "loss": 1.0125, "step": 910000 }, { "epoch": 26.73, "grad_norm": 4.370170593261719, "learning_rate": 3.2733851750885005e-05, "loss": 1.0134, "step": 910500 }, { "epoch": 26.75, "grad_norm": 1.755359172821045, "learning_rate": 3.2586759759553626e-05, "loss": 1.0174, "step": 911000 }, { "epoch": 26.76, "grad_norm": 2.099271774291992, "learning_rate": 3.243966776822225e-05, "loss": 1.0142, "step": 911500 }, { "epoch": 26.78, "grad_norm": 1.560284972190857, "learning_rate": 3.229257577689087e-05, "loss": 1.0393, "step": 912000 }, { "epoch": 26.79, "grad_norm": 2.1582424640655518, "learning_rate": 3.214548378555949e-05, "loss": 1.016, "step": 912500 }, { "epoch": 26.81, "grad_norm": 2.342576026916504, "learning_rate": 3.199839179422811e-05, "loss": 1.0262, "step": 913000 }, { "epoch": 26.82, "grad_norm": 1.4274934530258179, "learning_rate": 3.185129980289673e-05, "loss": 0.9991, "step": 913500 }, { "epoch": 26.84, "grad_norm": 3.858720064163208, "learning_rate": 3.170420781156535e-05, "loss": 1.0171, "step": 914000 }, { "epoch": 26.85, "grad_norm": 2.058018922805786, "learning_rate": 3.1557115820233974e-05, "loss": 1.0263, "step": 914500 }, { "epoch": 26.86, "grad_norm": 2.6626269817352295, "learning_rate": 3.1410023828902595e-05, "loss": 1.0257, "step": 915000 }, { "epoch": 26.88, "grad_norm": 5.859118938446045, "learning_rate": 3.1262931837571216e-05, "loss": 1.023, "step": 915500 }, { "epoch": 26.89, "grad_norm": 6.2187957763671875, "learning_rate": 3.111583984623984e-05, "loss": 1.0107, "step": 916000 }, { "epoch": 26.91, "grad_norm": 2.6373274326324463, "learning_rate": 3.096874785490846e-05, "loss": 1.0172, "step": 916500 }, { "epoch": 26.92, "grad_norm": 3.272045850753784, "learning_rate": 3.082165586357708e-05, "loss": 1.0175, "step": 917000 }, { "epoch": 26.94, "grad_norm": 1.6807574033737183, "learning_rate": 3.06745638722457e-05, "loss": 1.0295, "step": 917500 }, { "epoch": 26.95, "grad_norm": 4.664640426635742, "learning_rate": 3.052747188091432e-05, "loss": 1.0368, "step": 918000 }, { "epoch": 26.97, "grad_norm": 1.5610125064849854, "learning_rate": 3.0380379889582943e-05, "loss": 1.0365, "step": 918500 }, { "epoch": 26.98, "grad_norm": 2.0391860008239746, "learning_rate": 3.0233287898251564e-05, "loss": 1.0289, "step": 919000 }, { "epoch": 27.0, "grad_norm": 47.55400848388672, "learning_rate": 3.0086195906920185e-05, "loss": 1.0292, "step": 919500 }, { "epoch": 27.01, "grad_norm": 2.6204073429107666, "learning_rate": 2.9939103915588807e-05, "loss": 1.0069, "step": 920000 }, { "epoch": 27.03, "grad_norm": 9.747673034667969, "learning_rate": 2.9792011924257428e-05, "loss": 0.9997, "step": 920500 }, { "epoch": 27.04, "grad_norm": 2.057465076446533, "learning_rate": 2.9644919932926052e-05, "loss": 0.9993, "step": 921000 }, { "epoch": 27.06, "grad_norm": 8.264533996582031, "learning_rate": 2.9497827941594673e-05, "loss": 0.9862, "step": 921500 }, { "epoch": 27.07, "grad_norm": 4.143492698669434, "learning_rate": 2.9350735950263294e-05, "loss": 0.9774, "step": 922000 }, { "epoch": 27.08, "grad_norm": 2.2518653869628906, "learning_rate": 2.9203643958931916e-05, "loss": 1.0113, "step": 922500 }, { "epoch": 27.1, "grad_norm": 2.745732307434082, "learning_rate": 2.9056551967600537e-05, "loss": 0.9734, "step": 923000 }, { "epoch": 27.11, "grad_norm": 1.3990339040756226, "learning_rate": 2.8909459976269158e-05, "loss": 0.9918, "step": 923500 }, { "epoch": 27.13, "grad_norm": 2.024524211883545, "learning_rate": 2.876236798493778e-05, "loss": 1.0033, "step": 924000 }, { "epoch": 27.14, "grad_norm": 2.5025253295898438, "learning_rate": 2.86152759936064e-05, "loss": 1.0004, "step": 924500 }, { "epoch": 27.16, "grad_norm": 2.7478108406066895, "learning_rate": 2.846818400227502e-05, "loss": 0.9901, "step": 925000 }, { "epoch": 27.17, "grad_norm": 1.7393391132354736, "learning_rate": 2.8321092010943642e-05, "loss": 1.0035, "step": 925500 }, { "epoch": 27.19, "grad_norm": 1.3920613527297974, "learning_rate": 2.8174000019612263e-05, "loss": 1.0194, "step": 926000 }, { "epoch": 27.2, "grad_norm": 2.2148451805114746, "learning_rate": 2.8026908028280885e-05, "loss": 0.9858, "step": 926500 }, { "epoch": 27.22, "grad_norm": 1.7391964197158813, "learning_rate": 2.7879816036949506e-05, "loss": 1.0069, "step": 927000 }, { "epoch": 27.23, "grad_norm": 2.359649419784546, "learning_rate": 2.7732724045618127e-05, "loss": 0.9989, "step": 927500 }, { "epoch": 27.25, "grad_norm": 2.8471827507019043, "learning_rate": 2.7585632054286748e-05, "loss": 0.9832, "step": 928000 }, { "epoch": 27.26, "grad_norm": 1.788493037223816, "learning_rate": 2.743854006295537e-05, "loss": 1.0104, "step": 928500 }, { "epoch": 27.28, "grad_norm": 1.7663925886154175, "learning_rate": 2.7291448071623994e-05, "loss": 1.0031, "step": 929000 }, { "epoch": 27.29, "grad_norm": 2.699171304702759, "learning_rate": 2.7144356080292615e-05, "loss": 1.0037, "step": 929500 }, { "epoch": 27.31, "grad_norm": 12.16844367980957, "learning_rate": 2.6997264088961236e-05, "loss": 1.0109, "step": 930000 }, { "epoch": 27.32, "grad_norm": 2.0973167419433594, "learning_rate": 2.6850172097629857e-05, "loss": 1.0123, "step": 930500 }, { "epoch": 27.33, "grad_norm": 1.7330328226089478, "learning_rate": 2.6703080106298478e-05, "loss": 1.0055, "step": 931000 }, { "epoch": 27.35, "grad_norm": 5.502919673919678, "learning_rate": 2.65559881149671e-05, "loss": 1.003, "step": 931500 }, { "epoch": 27.36, "grad_norm": 2.3308217525482178, "learning_rate": 2.640889612363572e-05, "loss": 1.0059, "step": 932000 }, { "epoch": 27.38, "grad_norm": 1.6248698234558105, "learning_rate": 2.626180413230434e-05, "loss": 0.9849, "step": 932500 }, { "epoch": 27.39, "grad_norm": 3.2510266304016113, "learning_rate": 2.6114712140972963e-05, "loss": 0.9862, "step": 933000 }, { "epoch": 27.41, "grad_norm": 2.0555245876312256, "learning_rate": 2.5967620149641584e-05, "loss": 0.9844, "step": 933500 }, { "epoch": 27.42, "grad_norm": 7.706765651702881, "learning_rate": 2.5820528158310205e-05, "loss": 0.9913, "step": 934000 }, { "epoch": 27.44, "grad_norm": 1.9011601209640503, "learning_rate": 2.5673436166978826e-05, "loss": 1.0102, "step": 934500 }, { "epoch": 27.45, "grad_norm": 6.408154010772705, "learning_rate": 2.5526344175647447e-05, "loss": 0.996, "step": 935000 }, { "epoch": 27.47, "grad_norm": 6.746776580810547, "learning_rate": 2.5379252184316068e-05, "loss": 1.0273, "step": 935500 }, { "epoch": 27.48, "grad_norm": 33.82996368408203, "learning_rate": 2.523216019298469e-05, "loss": 1.007, "step": 936000 }, { "epoch": 27.5, "grad_norm": 1.6131765842437744, "learning_rate": 2.508506820165331e-05, "loss": 0.9855, "step": 936500 }, { "epoch": 27.51, "grad_norm": 2.5696966648101807, "learning_rate": 2.493797621032193e-05, "loss": 1.0153, "step": 937000 }, { "epoch": 27.53, "grad_norm": 1.5355910062789917, "learning_rate": 2.4790884218990556e-05, "loss": 0.9714, "step": 937500 }, { "epoch": 27.54, "grad_norm": 2.0435054302215576, "learning_rate": 2.4643792227659177e-05, "loss": 0.9997, "step": 938000 }, { "epoch": 27.55, "grad_norm": 1.6511478424072266, "learning_rate": 2.44967002363278e-05, "loss": 0.9865, "step": 938500 }, { "epoch": 27.57, "grad_norm": 1.9227514266967773, "learning_rate": 2.434960824499642e-05, "loss": 1.0103, "step": 939000 }, { "epoch": 27.58, "grad_norm": 16.868221282958984, "learning_rate": 2.420251625366504e-05, "loss": 0.9854, "step": 939500 }, { "epoch": 27.6, "grad_norm": 1.7769354581832886, "learning_rate": 2.4055424262333662e-05, "loss": 1.0176, "step": 940000 }, { "epoch": 27.61, "grad_norm": 1.6719156503677368, "learning_rate": 2.3908332271002283e-05, "loss": 0.9951, "step": 940500 }, { "epoch": 27.63, "grad_norm": 3.5106630325317383, "learning_rate": 2.3761240279670904e-05, "loss": 0.994, "step": 941000 }, { "epoch": 27.64, "grad_norm": 3.102487564086914, "learning_rate": 2.3614148288339525e-05, "loss": 0.9802, "step": 941500 }, { "epoch": 27.66, "grad_norm": 12.020054817199707, "learning_rate": 2.3467056297008146e-05, "loss": 0.9997, "step": 942000 }, { "epoch": 27.67, "grad_norm": 2.8283021450042725, "learning_rate": 2.3319964305676767e-05, "loss": 0.9892, "step": 942500 }, { "epoch": 27.69, "grad_norm": 37.08485412597656, "learning_rate": 2.317287231434539e-05, "loss": 0.9969, "step": 943000 }, { "epoch": 27.7, "grad_norm": 1.4365429878234863, "learning_rate": 2.302578032301401e-05, "loss": 0.9748, "step": 943500 }, { "epoch": 27.72, "grad_norm": 1.8424559831619263, "learning_rate": 2.287868833168263e-05, "loss": 0.9741, "step": 944000 }, { "epoch": 27.73, "grad_norm": 2.4849884510040283, "learning_rate": 2.2731596340351252e-05, "loss": 0.9896, "step": 944500 }, { "epoch": 27.75, "grad_norm": 2.0955851078033447, "learning_rate": 2.2584504349019873e-05, "loss": 1.012, "step": 945000 }, { "epoch": 27.76, "grad_norm": 5.421468257904053, "learning_rate": 2.2437412357688498e-05, "loss": 0.9681, "step": 945500 }, { "epoch": 27.77, "grad_norm": 3.2725253105163574, "learning_rate": 2.229032036635712e-05, "loss": 1.0059, "step": 946000 }, { "epoch": 27.79, "grad_norm": 2.741119861602783, "learning_rate": 2.214322837502574e-05, "loss": 0.9933, "step": 946500 }, { "epoch": 27.8, "grad_norm": 1.600335717201233, "learning_rate": 2.199613638369436e-05, "loss": 0.997, "step": 947000 }, { "epoch": 27.82, "grad_norm": 3.4619009494781494, "learning_rate": 2.1849044392362982e-05, "loss": 0.9976, "step": 947500 }, { "epoch": 27.83, "grad_norm": 10.748187065124512, "learning_rate": 2.1701952401031603e-05, "loss": 0.9727, "step": 948000 }, { "epoch": 27.85, "grad_norm": 3.452791690826416, "learning_rate": 2.1554860409700224e-05, "loss": 1.001, "step": 948500 }, { "epoch": 27.86, "grad_norm": 1.3464235067367554, "learning_rate": 2.1407768418368845e-05, "loss": 0.9876, "step": 949000 }, { "epoch": 27.88, "grad_norm": 1.6416382789611816, "learning_rate": 2.1260676427037467e-05, "loss": 1.0062, "step": 949500 }, { "epoch": 27.89, "grad_norm": 1.434910774230957, "learning_rate": 2.1113584435706088e-05, "loss": 0.9934, "step": 950000 }, { "epoch": 27.91, "grad_norm": 5.8801045417785645, "learning_rate": 2.096649244437471e-05, "loss": 0.9899, "step": 950500 }, { "epoch": 27.92, "grad_norm": 1.864310622215271, "learning_rate": 2.081940045304333e-05, "loss": 1.013, "step": 951000 }, { "epoch": 27.94, "grad_norm": 2.070906639099121, "learning_rate": 2.067230846171195e-05, "loss": 0.9947, "step": 951500 }, { "epoch": 27.95, "grad_norm": 2.5215396881103516, "learning_rate": 2.0525216470380572e-05, "loss": 0.984, "step": 952000 }, { "epoch": 27.97, "grad_norm": 1.2144149541854858, "learning_rate": 2.0378124479049193e-05, "loss": 0.9993, "step": 952500 }, { "epoch": 27.98, "grad_norm": 2.4629998207092285, "learning_rate": 2.0231032487717814e-05, "loss": 1.0146, "step": 953000 }, { "epoch": 28.0, "grad_norm": 4.50878381729126, "learning_rate": 2.008394049638644e-05, "loss": 1.0005, "step": 953500 }, { "epoch": 28.01, "grad_norm": 2.623155355453491, "learning_rate": 1.993684850505506e-05, "loss": 0.9928, "step": 954000 }, { "epoch": 28.02, "grad_norm": 2.886497735977173, "learning_rate": 1.978975651372368e-05, "loss": 0.9886, "step": 954500 }, { "epoch": 28.04, "grad_norm": 2.8885116577148438, "learning_rate": 1.9642664522392302e-05, "loss": 0.9788, "step": 955000 }, { "epoch": 28.05, "grad_norm": 1.946003794670105, "learning_rate": 1.9495572531060923e-05, "loss": 0.9647, "step": 955500 }, { "epoch": 28.07, "grad_norm": 12.426846504211426, "learning_rate": 1.9348480539729545e-05, "loss": 0.9753, "step": 956000 }, { "epoch": 28.08, "grad_norm": 1.905735969543457, "learning_rate": 1.9201388548398166e-05, "loss": 0.9855, "step": 956500 }, { "epoch": 28.1, "grad_norm": 2.33889102935791, "learning_rate": 1.9054296557066787e-05, "loss": 0.9744, "step": 957000 }, { "epoch": 28.11, "grad_norm": 2.3063101768493652, "learning_rate": 1.8907204565735408e-05, "loss": 0.9702, "step": 957500 }, { "epoch": 28.13, "grad_norm": 2.292107582092285, "learning_rate": 1.876011257440403e-05, "loss": 0.9888, "step": 958000 }, { "epoch": 28.14, "grad_norm": 2.144972324371338, "learning_rate": 1.8613020583072654e-05, "loss": 0.9754, "step": 958500 }, { "epoch": 28.16, "grad_norm": 1.5302867889404297, "learning_rate": 1.8465928591741275e-05, "loss": 0.9816, "step": 959000 }, { "epoch": 28.17, "grad_norm": 1.6462048292160034, "learning_rate": 1.8318836600409896e-05, "loss": 0.9403, "step": 959500 }, { "epoch": 28.19, "grad_norm": 2.6783134937286377, "learning_rate": 1.8171744609078517e-05, "loss": 0.9762, "step": 960000 }, { "epoch": 28.2, "grad_norm": 2.0703890323638916, "learning_rate": 1.8024652617747138e-05, "loss": 0.9756, "step": 960500 }, { "epoch": 28.22, "grad_norm": 2.4419829845428467, "learning_rate": 1.787756062641576e-05, "loss": 0.9684, "step": 961000 }, { "epoch": 28.23, "grad_norm": 4.775966644287109, "learning_rate": 1.773046863508438e-05, "loss": 0.9912, "step": 961500 }, { "epoch": 28.24, "grad_norm": 1.0155693292617798, "learning_rate": 1.7583376643753e-05, "loss": 0.97, "step": 962000 }, { "epoch": 28.26, "grad_norm": 2.105178117752075, "learning_rate": 1.7436284652421623e-05, "loss": 0.9699, "step": 962500 }, { "epoch": 28.27, "grad_norm": 11.809367179870605, "learning_rate": 1.7289192661090244e-05, "loss": 0.9869, "step": 963000 }, { "epoch": 28.29, "grad_norm": 2.5389130115509033, "learning_rate": 1.7142100669758865e-05, "loss": 0.9807, "step": 963500 }, { "epoch": 28.3, "grad_norm": 2.3344273567199707, "learning_rate": 1.6995008678427486e-05, "loss": 0.9861, "step": 964000 }, { "epoch": 28.32, "grad_norm": 2.3741703033447266, "learning_rate": 1.6847916687096107e-05, "loss": 0.9592, "step": 964500 }, { "epoch": 28.33, "grad_norm": 5.04816198348999, "learning_rate": 1.670082469576473e-05, "loss": 0.9903, "step": 965000 }, { "epoch": 28.35, "grad_norm": 1.2963931560516357, "learning_rate": 1.6553732704433353e-05, "loss": 0.9676, "step": 965500 }, { "epoch": 28.36, "grad_norm": 3.1595191955566406, "learning_rate": 1.6406640713101974e-05, "loss": 0.9942, "step": 966000 }, { "epoch": 28.38, "grad_norm": 1.5514929294586182, "learning_rate": 1.6259548721770595e-05, "loss": 0.9741, "step": 966500 }, { "epoch": 28.39, "grad_norm": 1.9268239736557007, "learning_rate": 1.6112456730439216e-05, "loss": 0.9869, "step": 967000 }, { "epoch": 28.41, "grad_norm": 2.7622382640838623, "learning_rate": 1.5965364739107837e-05, "loss": 0.9598, "step": 967500 }, { "epoch": 28.42, "grad_norm": 1.418778657913208, "learning_rate": 1.581827274777646e-05, "loss": 0.9464, "step": 968000 }, { "epoch": 28.44, "grad_norm": 2.1014456748962402, "learning_rate": 1.567118075644508e-05, "loss": 0.9846, "step": 968500 }, { "epoch": 28.45, "grad_norm": 14.58630657196045, "learning_rate": 1.55240887651137e-05, "loss": 0.9695, "step": 969000 }, { "epoch": 28.46, "grad_norm": 1.556394100189209, "learning_rate": 1.5376996773782322e-05, "loss": 0.9916, "step": 969500 }, { "epoch": 28.48, "grad_norm": 1.8601114749908447, "learning_rate": 1.5229904782450943e-05, "loss": 0.9552, "step": 970000 }, { "epoch": 28.49, "grad_norm": 2.2888152599334717, "learning_rate": 1.5082812791119564e-05, "loss": 0.972, "step": 970500 }, { "epoch": 28.51, "grad_norm": 2.9334394931793213, "learning_rate": 1.4935720799788187e-05, "loss": 0.9986, "step": 971000 }, { "epoch": 28.52, "grad_norm": 2.88075590133667, "learning_rate": 1.4788628808456808e-05, "loss": 0.9935, "step": 971500 }, { "epoch": 28.54, "grad_norm": 4.609189510345459, "learning_rate": 1.4641536817125429e-05, "loss": 0.9715, "step": 972000 }, { "epoch": 28.55, "grad_norm": 1.6899704933166504, "learning_rate": 1.449444482579405e-05, "loss": 0.9737, "step": 972500 }, { "epoch": 28.57, "grad_norm": 18.26349449157715, "learning_rate": 1.4347352834462671e-05, "loss": 0.9728, "step": 973000 }, { "epoch": 28.58, "grad_norm": 5.080548286437988, "learning_rate": 1.4200260843131292e-05, "loss": 0.9819, "step": 973500 }, { "epoch": 28.6, "grad_norm": 3.9338316917419434, "learning_rate": 1.4053168851799914e-05, "loss": 0.9758, "step": 974000 }, { "epoch": 28.61, "grad_norm": 1.4379346370697021, "learning_rate": 1.3906076860468535e-05, "loss": 0.9634, "step": 974500 }, { "epoch": 28.63, "grad_norm": 2.943136215209961, "learning_rate": 1.3758984869137157e-05, "loss": 0.9519, "step": 975000 }, { "epoch": 28.64, "grad_norm": 3.1957294940948486, "learning_rate": 1.3611892877805779e-05, "loss": 0.9727, "step": 975500 }, { "epoch": 28.66, "grad_norm": 2.38423752784729, "learning_rate": 1.34648008864744e-05, "loss": 0.9787, "step": 976000 }, { "epoch": 28.67, "grad_norm": 2.059079885482788, "learning_rate": 1.331770889514302e-05, "loss": 0.9851, "step": 976500 }, { "epoch": 28.69, "grad_norm": 1.8208869695663452, "learning_rate": 1.3170616903811642e-05, "loss": 0.9665, "step": 977000 }, { "epoch": 28.7, "grad_norm": 2.108597755432129, "learning_rate": 1.3023524912480263e-05, "loss": 0.9941, "step": 977500 }, { "epoch": 28.71, "grad_norm": 1.5259730815887451, "learning_rate": 1.2876432921148884e-05, "loss": 0.98, "step": 978000 }, { "epoch": 28.73, "grad_norm": 3.095134735107422, "learning_rate": 1.2729340929817505e-05, "loss": 0.9962, "step": 978500 }, { "epoch": 28.74, "grad_norm": 1.49867582321167, "learning_rate": 1.2582248938486128e-05, "loss": 0.9714, "step": 979000 }, { "epoch": 28.76, "grad_norm": 1.5933908224105835, "learning_rate": 1.243515694715475e-05, "loss": 0.999, "step": 979500 }, { "epoch": 28.77, "grad_norm": 2.663355588912964, "learning_rate": 1.228806495582337e-05, "loss": 0.9682, "step": 980000 }, { "epoch": 28.79, "grad_norm": 104.24024200439453, "learning_rate": 1.2140972964491992e-05, "loss": 0.9774, "step": 980500 }, { "epoch": 28.8, "grad_norm": 1.7283943891525269, "learning_rate": 1.1993880973160613e-05, "loss": 0.9607, "step": 981000 }, { "epoch": 28.82, "grad_norm": 2.0852441787719727, "learning_rate": 1.1846788981829234e-05, "loss": 1.0005, "step": 981500 }, { "epoch": 28.83, "grad_norm": 2.0442757606506348, "learning_rate": 1.1699696990497855e-05, "loss": 0.9836, "step": 982000 }, { "epoch": 28.85, "grad_norm": 1.5879892110824585, "learning_rate": 1.1552604999166476e-05, "loss": 0.9973, "step": 982500 }, { "epoch": 28.86, "grad_norm": 2.7892065048217773, "learning_rate": 1.14055130078351e-05, "loss": 0.9615, "step": 983000 }, { "epoch": 28.88, "grad_norm": 3.826296329498291, "learning_rate": 1.1258421016503722e-05, "loss": 0.9746, "step": 983500 }, { "epoch": 28.89, "grad_norm": 7.535625457763672, "learning_rate": 1.1111329025172343e-05, "loss": 0.9685, "step": 984000 }, { "epoch": 28.91, "grad_norm": 1.7329784631729126, "learning_rate": 1.0964237033840964e-05, "loss": 0.9745, "step": 984500 }, { "epoch": 28.92, "grad_norm": 1.9252818822860718, "learning_rate": 1.0817145042509585e-05, "loss": 0.9693, "step": 985000 }, { "epoch": 28.93, "grad_norm": 2.0630569458007812, "learning_rate": 1.0670053051178206e-05, "loss": 0.9634, "step": 985500 }, { "epoch": 28.95, "grad_norm": 1.6531202793121338, "learning_rate": 1.0522961059846827e-05, "loss": 0.9702, "step": 986000 }, { "epoch": 28.96, "grad_norm": 32.64999771118164, "learning_rate": 1.037586906851545e-05, "loss": 0.9593, "step": 986500 }, { "epoch": 28.98, "grad_norm": 5.578742980957031, "learning_rate": 1.0228777077184071e-05, "loss": 0.9624, "step": 987000 }, { "epoch": 28.99, "grad_norm": 1.9109100103378296, "learning_rate": 1.0081685085852692e-05, "loss": 0.9698, "step": 987500 }, { "epoch": 29.01, "grad_norm": 3.434389352798462, "learning_rate": 9.934593094521313e-06, "loss": 0.9824, "step": 988000 }, { "epoch": 29.02, "grad_norm": 3.804184913635254, "learning_rate": 9.787501103189935e-06, "loss": 0.9534, "step": 988500 }, { "epoch": 29.04, "grad_norm": 2.4718267917633057, "learning_rate": 9.640409111858556e-06, "loss": 0.9533, "step": 989000 }, { "epoch": 29.05, "grad_norm": 2.81648588180542, "learning_rate": 9.493317120527177e-06, "loss": 0.9688, "step": 989500 }, { "epoch": 29.07, "grad_norm": 3.936281681060791, "learning_rate": 9.346225129195798e-06, "loss": 0.9486, "step": 990000 }, { "epoch": 29.08, "grad_norm": 4.693751335144043, "learning_rate": 9.19913313786442e-06, "loss": 0.9505, "step": 990500 }, { "epoch": 29.1, "grad_norm": 3.5278286933898926, "learning_rate": 9.052041146533042e-06, "loss": 0.9338, "step": 991000 }, { "epoch": 29.11, "grad_norm": 2.9467105865478516, "learning_rate": 8.904949155201663e-06, "loss": 0.9602, "step": 991500 }, { "epoch": 29.13, "grad_norm": 2.929056406021118, "learning_rate": 8.757857163870284e-06, "loss": 0.9469, "step": 992000 }, { "epoch": 29.14, "grad_norm": 4.979098320007324, "learning_rate": 8.610765172538905e-06, "loss": 0.9364, "step": 992500 }, { "epoch": 29.15, "grad_norm": 3.7940807342529297, "learning_rate": 8.463673181207526e-06, "loss": 0.9407, "step": 993000 }, { "epoch": 29.17, "grad_norm": 2.5154898166656494, "learning_rate": 8.316581189876148e-06, "loss": 0.9504, "step": 993500 }, { "epoch": 29.18, "grad_norm": 3.7596309185028076, "learning_rate": 8.169489198544769e-06, "loss": 0.9762, "step": 994000 }, { "epoch": 29.2, "grad_norm": 2.13797926902771, "learning_rate": 8.022397207213392e-06, "loss": 0.9605, "step": 994500 }, { "epoch": 29.21, "grad_norm": 3.1647770404815674, "learning_rate": 7.875305215882013e-06, "loss": 0.9611, "step": 995000 }, { "epoch": 29.23, "grad_norm": 3.417736768722534, "learning_rate": 7.728213224550634e-06, "loss": 0.9611, "step": 995500 }, { "epoch": 29.24, "grad_norm": 4.05425500869751, "learning_rate": 7.581121233219255e-06, "loss": 0.9733, "step": 996000 }, { "epoch": 29.26, "grad_norm": 2.117457866668701, "learning_rate": 7.434029241887876e-06, "loss": 0.9494, "step": 996500 }, { "epoch": 29.27, "grad_norm": 9.484780311584473, "learning_rate": 7.286937250556497e-06, "loss": 0.9842, "step": 997000 }, { "epoch": 29.29, "grad_norm": 5.586018085479736, "learning_rate": 7.139845259225119e-06, "loss": 0.955, "step": 997500 }, { "epoch": 29.3, "grad_norm": 6.439801216125488, "learning_rate": 6.99275326789374e-06, "loss": 0.9608, "step": 998000 }, { "epoch": 29.32, "grad_norm": 2.890882730484009, "learning_rate": 6.845661276562361e-06, "loss": 0.9649, "step": 998500 }, { "epoch": 29.33, "grad_norm": 2.0991222858428955, "learning_rate": 6.6985692852309825e-06, "loss": 0.9476, "step": 999000 }, { "epoch": 29.35, "grad_norm": 2.40987229347229, "learning_rate": 6.5514772938996045e-06, "loss": 0.9537, "step": 999500 }, { "epoch": 29.36, "grad_norm": 2.95318865776062, "learning_rate": 6.4043853025682256e-06, "loss": 0.9749, "step": 1000000 }, { "epoch": 29.38, "grad_norm": 5.403099060058594, "learning_rate": 6.257293311236847e-06, "loss": 0.9496, "step": 1000500 }, { "epoch": 29.39, "grad_norm": 3.3474481105804443, "learning_rate": 6.110201319905468e-06, "loss": 0.9756, "step": 1001000 }, { "epoch": 29.4, "grad_norm": 2.7617027759552, "learning_rate": 5.96310932857409e-06, "loss": 0.944, "step": 1001500 }, { "epoch": 29.42, "grad_norm": 3.7849531173706055, "learning_rate": 5.816017337242711e-06, "loss": 0.9616, "step": 1002000 }, { "epoch": 29.43, "grad_norm": 2.1899170875549316, "learning_rate": 5.668925345911332e-06, "loss": 0.9578, "step": 1002500 }, { "epoch": 29.45, "grad_norm": 2.423264265060425, "learning_rate": 5.521833354579953e-06, "loss": 0.9886, "step": 1003000 }, { "epoch": 29.46, "grad_norm": 10.377019882202148, "learning_rate": 5.374741363248575e-06, "loss": 0.9559, "step": 1003500 }, { "epoch": 29.48, "grad_norm": 3.244654655456543, "learning_rate": 5.227649371917196e-06, "loss": 0.9467, "step": 1004000 }, { "epoch": 29.49, "grad_norm": 3.0262067317962646, "learning_rate": 5.080557380585817e-06, "loss": 0.9825, "step": 1004500 }, { "epoch": 29.51, "grad_norm": 1.8769322633743286, "learning_rate": 4.933465389254439e-06, "loss": 0.9835, "step": 1005000 }, { "epoch": 29.52, "grad_norm": 3.122030735015869, "learning_rate": 4.786373397923061e-06, "loss": 0.9686, "step": 1005500 }, { "epoch": 29.54, "grad_norm": 3.3955576419830322, "learning_rate": 4.639281406591682e-06, "loss": 0.9549, "step": 1006000 }, { "epoch": 29.55, "grad_norm": 11.812577247619629, "learning_rate": 4.492189415260304e-06, "loss": 0.944, "step": 1006500 }, { "epoch": 29.57, "grad_norm": 3.2719273567199707, "learning_rate": 4.345097423928925e-06, "loss": 0.9675, "step": 1007000 }, { "epoch": 29.58, "grad_norm": 2.0601563453674316, "learning_rate": 4.198005432597546e-06, "loss": 0.9816, "step": 1007500 }, { "epoch": 29.6, "grad_norm": 2.1465187072753906, "learning_rate": 4.050913441266167e-06, "loss": 0.9386, "step": 1008000 }, { "epoch": 29.61, "grad_norm": 3.070183753967285, "learning_rate": 3.903821449934789e-06, "loss": 0.9729, "step": 1008500 }, { "epoch": 29.62, "grad_norm": 8.74717903137207, "learning_rate": 3.7567294586034105e-06, "loss": 0.9483, "step": 1009000 }, { "epoch": 29.64, "grad_norm": 4.731165409088135, "learning_rate": 3.6096374672720316e-06, "loss": 0.9699, "step": 1009500 }, { "epoch": 29.65, "grad_norm": 3.4259703159332275, "learning_rate": 3.462545475940653e-06, "loss": 0.9768, "step": 1010000 }, { "epoch": 29.67, "grad_norm": 2.1280946731567383, "learning_rate": 3.3154534846092743e-06, "loss": 0.9732, "step": 1010500 }, { "epoch": 29.68, "grad_norm": 2.901146411895752, "learning_rate": 3.168361493277896e-06, "loss": 0.968, "step": 1011000 }, { "epoch": 29.7, "grad_norm": 3.9223461151123047, "learning_rate": 3.021269501946517e-06, "loss": 0.9499, "step": 1011500 }, { "epoch": 29.71, "grad_norm": 2.9555001258850098, "learning_rate": 2.8741775106151385e-06, "loss": 0.9306, "step": 1012000 }, { "epoch": 29.73, "grad_norm": 3.057680606842041, "learning_rate": 2.7270855192837596e-06, "loss": 0.9558, "step": 1012500 }, { "epoch": 29.74, "grad_norm": 2.978799343109131, "learning_rate": 2.579993527952381e-06, "loss": 0.9701, "step": 1013000 }, { "epoch": 29.76, "grad_norm": 2.0970911979675293, "learning_rate": 2.4329015366210023e-06, "loss": 0.9577, "step": 1013500 }, { "epoch": 29.77, "grad_norm": 2.8144404888153076, "learning_rate": 2.285809545289624e-06, "loss": 0.9587, "step": 1014000 }, { "epoch": 29.79, "grad_norm": 2.756420135498047, "learning_rate": 2.1387175539582454e-06, "loss": 0.9554, "step": 1014500 }, { "epoch": 29.8, "grad_norm": 3.5521059036254883, "learning_rate": 1.9916255626268665e-06, "loss": 0.95, "step": 1015000 }, { "epoch": 29.82, "grad_norm": 34.70965576171875, "learning_rate": 1.844533571295488e-06, "loss": 0.9708, "step": 1015500 }, { "epoch": 29.83, "grad_norm": 3.2949025630950928, "learning_rate": 1.6974415799641094e-06, "loss": 0.9673, "step": 1016000 }, { "epoch": 29.84, "grad_norm": 2.9649112224578857, "learning_rate": 1.5503495886327307e-06, "loss": 0.9758, "step": 1016500 }, { "epoch": 29.86, "grad_norm": 2.4358630180358887, "learning_rate": 1.403257597301352e-06, "loss": 0.9793, "step": 1017000 }, { "epoch": 29.87, "grad_norm": 3.3332042694091797, "learning_rate": 1.2561656059699736e-06, "loss": 0.9783, "step": 1017500 }, { "epoch": 29.89, "grad_norm": 5.788940906524658, "learning_rate": 1.1090736146385948e-06, "loss": 0.9519, "step": 1018000 }, { "epoch": 29.9, "grad_norm": 2.6301000118255615, "learning_rate": 9.619816233072163e-07, "loss": 0.9558, "step": 1018500 }, { "epoch": 29.92, "grad_norm": 2.9536306858062744, "learning_rate": 8.148896319758376e-07, "loss": 0.9466, "step": 1019000 }, { "epoch": 29.93, "grad_norm": 2.9202470779418945, "learning_rate": 6.67797640644459e-07, "loss": 0.9695, "step": 1019500 }, { "epoch": 29.95, "grad_norm": 3.0824179649353027, "learning_rate": 5.207056493130803e-07, "loss": 0.9361, "step": 1020000 }, { "epoch": 29.96, "grad_norm": 2.5186848640441895, "learning_rate": 3.736136579817017e-07, "loss": 0.9476, "step": 1020500 }, { "epoch": 29.98, "grad_norm": 5.588615894317627, "learning_rate": 2.265216666503231e-07, "loss": 0.9431, "step": 1021000 }, { "epoch": 29.99, "grad_norm": 2.025895595550537, "learning_rate": 7.942967531894445e-08, "loss": 0.9567, "step": 1021500 } ], "logging_steps": 500, "max_steps": 1021770, "num_input_tokens_seen": 0, "num_train_epochs": 30, "save_steps": 500, "total_flos": 1.7792790799000776e+21, "train_batch_size": 2, "trial_name": null, "trial_params": null }