{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.995163584637269, "eval_steps": 50, "global_step": 987, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0030346135609293505, "grad_norm": 5.274968147277832, "learning_rate": 2.0202020202020205e-07, "loss": 0.8522, "step": 1 }, { "epoch": 0.006069227121858701, "grad_norm": 5.754838466644287, "learning_rate": 4.040404040404041e-07, "loss": 0.854, "step": 2 }, { "epoch": 0.009103840682788052, "grad_norm": 5.632802963256836, "learning_rate": 6.060606060606061e-07, "loss": 0.873, "step": 3 }, { "epoch": 0.012138454243717402, "grad_norm": 5.394160747528076, "learning_rate": 8.080808080808082e-07, "loss": 0.8563, "step": 4 }, { "epoch": 0.015173067804646752, "grad_norm": 4.5878424644470215, "learning_rate": 1.01010101010101e-06, "loss": 0.8425, "step": 5 }, { "epoch": 0.018207681365576104, "grad_norm": 4.280163764953613, "learning_rate": 1.2121212121212122e-06, "loss": 0.8569, "step": 6 }, { "epoch": 0.021242294926505454, "grad_norm": 3.5766515731811523, "learning_rate": 1.4141414141414143e-06, "loss": 0.8309, "step": 7 }, { "epoch": 0.024276908487434804, "grad_norm": 4.070709228515625, "learning_rate": 1.6161616161616164e-06, "loss": 0.825, "step": 8 }, { "epoch": 0.027311522048364154, "grad_norm": 4.395944595336914, "learning_rate": 1.8181818181818183e-06, "loss": 0.8419, "step": 9 }, { "epoch": 0.030346135609293504, "grad_norm": 3.765310764312744, "learning_rate": 2.02020202020202e-06, "loss": 0.8426, "step": 10 }, { "epoch": 0.03338074917022286, "grad_norm": 2.906996488571167, "learning_rate": 2.222222222222222e-06, "loss": 0.8263, "step": 11 }, { "epoch": 0.03641536273115221, "grad_norm": 2.8421216011047363, "learning_rate": 2.4242424242424244e-06, "loss": 0.8242, "step": 12 }, { "epoch": 0.03944997629208156, "grad_norm": 2.6818017959594727, "learning_rate": 2.6262626262626267e-06, "loss": 0.8206, "step": 13 }, { "epoch": 0.04248458985301091, "grad_norm": 2.8597841262817383, "learning_rate": 2.8282828282828286e-06, "loss": 0.8201, "step": 14 }, { "epoch": 0.04551920341394026, "grad_norm": 2.673666000366211, "learning_rate": 3.0303030303030305e-06, "loss": 0.8215, "step": 15 }, { "epoch": 0.04855381697486961, "grad_norm": 2.636281728744507, "learning_rate": 3.232323232323233e-06, "loss": 0.8037, "step": 16 }, { "epoch": 0.05158843053579896, "grad_norm": 2.502042770385742, "learning_rate": 3.4343434343434347e-06, "loss": 0.8162, "step": 17 }, { "epoch": 0.05462304409672831, "grad_norm": 2.562307834625244, "learning_rate": 3.6363636363636366e-06, "loss": 0.8066, "step": 18 }, { "epoch": 0.05765765765765766, "grad_norm": 2.6667816638946533, "learning_rate": 3.8383838383838385e-06, "loss": 0.8197, "step": 19 }, { "epoch": 0.06069227121858701, "grad_norm": 2.293628215789795, "learning_rate": 4.04040404040404e-06, "loss": 0.8092, "step": 20 }, { "epoch": 0.06372688477951635, "grad_norm": 2.4843485355377197, "learning_rate": 4.242424242424243e-06, "loss": 0.8082, "step": 21 }, { "epoch": 0.06676149834044572, "grad_norm": 2.5178024768829346, "learning_rate": 4.444444444444444e-06, "loss": 0.8165, "step": 22 }, { "epoch": 0.06979611190137507, "grad_norm": 2.6686830520629883, "learning_rate": 4.646464646464647e-06, "loss": 0.773, "step": 23 }, { "epoch": 0.07283072546230442, "grad_norm": 2.2187390327453613, "learning_rate": 4.848484848484849e-06, "loss": 0.8026, "step": 24 }, { "epoch": 0.07586533902323377, "grad_norm": 2.4837453365325928, "learning_rate": 5.0505050505050515e-06, "loss": 0.7962, "step": 25 }, { "epoch": 0.07889995258416312, "grad_norm": 2.3953895568847656, "learning_rate": 5.252525252525253e-06, "loss": 0.7902, "step": 26 }, { "epoch": 0.08193456614509247, "grad_norm": 2.384993553161621, "learning_rate": 5.4545454545454545e-06, "loss": 0.8133, "step": 27 }, { "epoch": 0.08496917970602182, "grad_norm": 2.5588831901550293, "learning_rate": 5.656565656565657e-06, "loss": 0.8043, "step": 28 }, { "epoch": 0.08800379326695117, "grad_norm": 2.254196882247925, "learning_rate": 5.858585858585859e-06, "loss": 0.7967, "step": 29 }, { "epoch": 0.09103840682788052, "grad_norm": 2.2686617374420166, "learning_rate": 6.060606060606061e-06, "loss": 0.8036, "step": 30 }, { "epoch": 0.09407302038880987, "grad_norm": 2.3785650730133057, "learning_rate": 6.262626262626264e-06, "loss": 0.799, "step": 31 }, { "epoch": 0.09710763394973922, "grad_norm": 2.4035139083862305, "learning_rate": 6.464646464646466e-06, "loss": 0.8086, "step": 32 }, { "epoch": 0.10014224751066857, "grad_norm": 2.2935521602630615, "learning_rate": 6.666666666666667e-06, "loss": 0.7842, "step": 33 }, { "epoch": 0.10317686107159792, "grad_norm": 2.3163795471191406, "learning_rate": 6.868686868686869e-06, "loss": 0.8041, "step": 34 }, { "epoch": 0.10621147463252727, "grad_norm": 2.3953793048858643, "learning_rate": 7.070707070707071e-06, "loss": 0.7929, "step": 35 }, { "epoch": 0.10924608819345662, "grad_norm": 2.4116768836975098, "learning_rate": 7.272727272727273e-06, "loss": 0.7943, "step": 36 }, { "epoch": 0.11228070175438597, "grad_norm": 2.5221264362335205, "learning_rate": 7.474747474747476e-06, "loss": 0.8036, "step": 37 }, { "epoch": 0.11531531531531532, "grad_norm": 2.4730491638183594, "learning_rate": 7.676767676767677e-06, "loss": 0.8063, "step": 38 }, { "epoch": 0.11834992887624467, "grad_norm": 2.1653268337249756, "learning_rate": 7.87878787878788e-06, "loss": 0.8009, "step": 39 }, { "epoch": 0.12138454243717402, "grad_norm": 2.266632080078125, "learning_rate": 8.08080808080808e-06, "loss": 0.7825, "step": 40 }, { "epoch": 0.12441915599810337, "grad_norm": 2.2630982398986816, "learning_rate": 8.282828282828283e-06, "loss": 0.8004, "step": 41 }, { "epoch": 0.1274537695590327, "grad_norm": 2.403125762939453, "learning_rate": 8.484848484848486e-06, "loss": 0.7911, "step": 42 }, { "epoch": 0.13048838311996205, "grad_norm": 2.4483842849731445, "learning_rate": 8.686868686868687e-06, "loss": 0.7861, "step": 43 }, { "epoch": 0.13352299668089143, "grad_norm": 2.346780776977539, "learning_rate": 8.888888888888888e-06, "loss": 0.7969, "step": 44 }, { "epoch": 0.13655761024182078, "grad_norm": 2.3152859210968018, "learning_rate": 9.090909090909091e-06, "loss": 0.7842, "step": 45 }, { "epoch": 0.13959222380275013, "grad_norm": 2.2333388328552246, "learning_rate": 9.292929292929294e-06, "loss": 0.7886, "step": 46 }, { "epoch": 0.14262683736367948, "grad_norm": 2.360898017883301, "learning_rate": 9.494949494949497e-06, "loss": 0.8146, "step": 47 }, { "epoch": 0.14566145092460883, "grad_norm": 2.5461983680725098, "learning_rate": 9.696969696969698e-06, "loss": 0.8005, "step": 48 }, { "epoch": 0.14869606448553818, "grad_norm": 2.4684877395629883, "learning_rate": 9.8989898989899e-06, "loss": 0.8031, "step": 49 }, { "epoch": 0.15173067804646753, "grad_norm": 2.3025074005126953, "learning_rate": 1.0101010101010103e-05, "loss": 0.7828, "step": 50 }, { "epoch": 0.15173067804646753, "eval_loss": NaN, "eval_runtime": 233.599, "eval_samples_per_second": 17.8, "eval_steps_per_second": 4.452, "step": 50 }, { "epoch": 0.15476529160739688, "grad_norm": 2.4276065826416016, "learning_rate": 1.0303030303030304e-05, "loss": 0.7801, "step": 51 }, { "epoch": 0.15779990516832623, "grad_norm": 2.279287815093994, "learning_rate": 1.0505050505050507e-05, "loss": 0.8055, "step": 52 }, { "epoch": 0.16083451872925558, "grad_norm": 2.7459404468536377, "learning_rate": 1.0707070707070708e-05, "loss": 0.7864, "step": 53 }, { "epoch": 0.16386913229018493, "grad_norm": 2.3488147258758545, "learning_rate": 1.0909090909090909e-05, "loss": 0.7869, "step": 54 }, { "epoch": 0.16690374585111428, "grad_norm": 2.547166109085083, "learning_rate": 1.1111111111111113e-05, "loss": 0.7869, "step": 55 }, { "epoch": 0.16993835941204363, "grad_norm": 2.4360105991363525, "learning_rate": 1.1313131313131314e-05, "loss": 0.7883, "step": 56 }, { "epoch": 0.17297297297297298, "grad_norm": 2.8864669799804688, "learning_rate": 1.1515151515151517e-05, "loss": 0.7927, "step": 57 }, { "epoch": 0.17600758653390233, "grad_norm": 2.3167998790740967, "learning_rate": 1.1717171717171718e-05, "loss": 0.7999, "step": 58 }, { "epoch": 0.17904220009483168, "grad_norm": 3.075058698654175, "learning_rate": 1.191919191919192e-05, "loss": 0.7888, "step": 59 }, { "epoch": 0.18207681365576103, "grad_norm": 2.38899827003479, "learning_rate": 1.2121212121212122e-05, "loss": 0.787, "step": 60 }, { "epoch": 0.18511142721669038, "grad_norm": 3.121044874191284, "learning_rate": 1.2323232323232323e-05, "loss": 0.8053, "step": 61 }, { "epoch": 0.18814604077761973, "grad_norm": 2.580725908279419, "learning_rate": 1.2525252525252527e-05, "loss": 0.7819, "step": 62 }, { "epoch": 0.19118065433854908, "grad_norm": 3.1028575897216797, "learning_rate": 1.2727272727272728e-05, "loss": 0.7886, "step": 63 }, { "epoch": 0.19421526789947843, "grad_norm": 2.6575424671173096, "learning_rate": 1.2929292929292931e-05, "loss": 0.7904, "step": 64 }, { "epoch": 0.19724988146040778, "grad_norm": 2.8595755100250244, "learning_rate": 1.3131313131313132e-05, "loss": 0.7955, "step": 65 }, { "epoch": 0.20028449502133713, "grad_norm": 2.235410451889038, "learning_rate": 1.3333333333333333e-05, "loss": 0.796, "step": 66 }, { "epoch": 0.20331910858226648, "grad_norm": 2.7857375144958496, "learning_rate": 1.3535353535353538e-05, "loss": 0.8008, "step": 67 }, { "epoch": 0.20635372214319583, "grad_norm": 2.310331106185913, "learning_rate": 1.3737373737373739e-05, "loss": 0.7802, "step": 68 }, { "epoch": 0.20938833570412518, "grad_norm": 2.8039300441741943, "learning_rate": 1.3939393939393942e-05, "loss": 0.7946, "step": 69 }, { "epoch": 0.21242294926505453, "grad_norm": 2.345369815826416, "learning_rate": 1.4141414141414143e-05, "loss": 0.7835, "step": 70 }, { "epoch": 0.21545756282598388, "grad_norm": 2.8384764194488525, "learning_rate": 1.4343434343434344e-05, "loss": 0.7855, "step": 71 }, { "epoch": 0.21849217638691323, "grad_norm": 2.371159076690674, "learning_rate": 1.4545454545454546e-05, "loss": 0.7689, "step": 72 }, { "epoch": 0.22152678994784258, "grad_norm": 2.9578475952148438, "learning_rate": 1.4747474747474747e-05, "loss": 0.7918, "step": 73 }, { "epoch": 0.22456140350877193, "grad_norm": 2.8900325298309326, "learning_rate": 1.4949494949494952e-05, "loss": 0.7746, "step": 74 }, { "epoch": 0.22759601706970128, "grad_norm": 2.4469008445739746, "learning_rate": 1.5151515151515153e-05, "loss": 0.7919, "step": 75 }, { "epoch": 0.23063063063063063, "grad_norm": 2.410154104232788, "learning_rate": 1.5353535353535354e-05, "loss": 0.7987, "step": 76 }, { "epoch": 0.23366524419155998, "grad_norm": 2.599086284637451, "learning_rate": 1.555555555555556e-05, "loss": 0.7714, "step": 77 }, { "epoch": 0.23669985775248933, "grad_norm": 2.4329092502593994, "learning_rate": 1.575757575757576e-05, "loss": 0.8041, "step": 78 }, { "epoch": 0.23973447131341868, "grad_norm": 2.7239413261413574, "learning_rate": 1.595959595959596e-05, "loss": 0.7941, "step": 79 }, { "epoch": 0.24276908487434803, "grad_norm": 2.8684301376342773, "learning_rate": 1.616161616161616e-05, "loss": 0.8076, "step": 80 }, { "epoch": 0.24580369843527738, "grad_norm": 3.004608631134033, "learning_rate": 1.6363636363636366e-05, "loss": 0.7893, "step": 81 }, { "epoch": 0.24883831199620673, "grad_norm": 2.6391353607177734, "learning_rate": 1.6565656565656567e-05, "loss": 0.8116, "step": 82 }, { "epoch": 0.2518729255571361, "grad_norm": 3.3293023109436035, "learning_rate": 1.6767676767676768e-05, "loss": 0.7952, "step": 83 }, { "epoch": 0.2549075391180654, "grad_norm": 2.5788087844848633, "learning_rate": 1.6969696969696972e-05, "loss": 0.7905, "step": 84 }, { "epoch": 0.2579421526789948, "grad_norm": 3.4907588958740234, "learning_rate": 1.7171717171717173e-05, "loss": 0.7903, "step": 85 }, { "epoch": 0.2609767662399241, "grad_norm": 2.8050403594970703, "learning_rate": 1.7373737373737375e-05, "loss": 0.7813, "step": 86 }, { "epoch": 0.2640113798008535, "grad_norm": 3.180819034576416, "learning_rate": 1.7575757575757576e-05, "loss": 0.7797, "step": 87 }, { "epoch": 0.26704599336178286, "grad_norm": 2.614614486694336, "learning_rate": 1.7777777777777777e-05, "loss": 0.7903, "step": 88 }, { "epoch": 0.2700806069227122, "grad_norm": 3.212218761444092, "learning_rate": 1.797979797979798e-05, "loss": 0.7927, "step": 89 }, { "epoch": 0.27311522048364156, "grad_norm": 2.6879336833953857, "learning_rate": 1.8181818181818182e-05, "loss": 0.8099, "step": 90 }, { "epoch": 0.2761498340445709, "grad_norm": 3.2700448036193848, "learning_rate": 1.8383838383838387e-05, "loss": 0.794, "step": 91 }, { "epoch": 0.27918444760550026, "grad_norm": 3.1249783039093018, "learning_rate": 1.8585858585858588e-05, "loss": 0.7807, "step": 92 }, { "epoch": 0.2822190611664296, "grad_norm": 2.4789459705352783, "learning_rate": 1.8787878787878792e-05, "loss": 0.7829, "step": 93 }, { "epoch": 0.28525367472735896, "grad_norm": 2.624887466430664, "learning_rate": 1.8989898989898993e-05, "loss": 0.8043, "step": 94 }, { "epoch": 0.2882882882882883, "grad_norm": 2.3336539268493652, "learning_rate": 1.9191919191919194e-05, "loss": 0.7827, "step": 95 }, { "epoch": 0.29132290184921766, "grad_norm": 2.5490119457244873, "learning_rate": 1.9393939393939395e-05, "loss": 0.8072, "step": 96 }, { "epoch": 0.294357515410147, "grad_norm": 2.5659990310668945, "learning_rate": 1.9595959595959596e-05, "loss": 0.7986, "step": 97 }, { "epoch": 0.29739212897107636, "grad_norm": 2.3554656505584717, "learning_rate": 1.97979797979798e-05, "loss": 0.8059, "step": 98 }, { "epoch": 0.3004267425320057, "grad_norm": 2.4364328384399414, "learning_rate": 2e-05, "loss": 0.7874, "step": 99 }, { "epoch": 0.30346135609293506, "grad_norm": 2.3030965328216553, "learning_rate": 1.9999937418875125e-05, "loss": 0.7854, "step": 100 }, { "epoch": 0.30346135609293506, "eval_loss": NaN, "eval_runtime": 204.7053, "eval_samples_per_second": 20.312, "eval_steps_per_second": 5.08, "step": 100 }, { "epoch": 0.3064959696538644, "grad_norm": 2.318124294281006, "learning_rate": 1.9999749676283775e-05, "loss": 0.7964, "step": 101 }, { "epoch": 0.30953058321479376, "grad_norm": 2.2906177043914795, "learning_rate": 1.9999436774575783e-05, "loss": 0.8049, "step": 102 }, { "epoch": 0.3125651967757231, "grad_norm": 2.2947778701782227, "learning_rate": 1.999899871766749e-05, "loss": 0.7881, "step": 103 }, { "epoch": 0.31559981033665246, "grad_norm": 2.548788547515869, "learning_rate": 1.999843551104172e-05, "loss": 0.803, "step": 104 }, { "epoch": 0.3186344238975818, "grad_norm": 2.3735954761505127, "learning_rate": 1.9997747161747696e-05, "loss": 0.8044, "step": 105 }, { "epoch": 0.32166903745851116, "grad_norm": 2.620364189147949, "learning_rate": 1.9996933678400948e-05, "loss": 0.7806, "step": 106 }, { "epoch": 0.3247036510194405, "grad_norm": 2.3897509574890137, "learning_rate": 1.999599507118322e-05, "loss": 0.7862, "step": 107 }, { "epoch": 0.32773826458036986, "grad_norm": 2.6068966388702393, "learning_rate": 1.9994931351842327e-05, "loss": 0.8051, "step": 108 }, { "epoch": 0.3307728781412992, "grad_norm": 2.560683488845825, "learning_rate": 1.999374253369202e-05, "loss": 0.7956, "step": 109 }, { "epoch": 0.33380749170222856, "grad_norm": 2.1629045009613037, "learning_rate": 1.999242863161182e-05, "loss": 0.8074, "step": 110 }, { "epoch": 0.3368421052631579, "grad_norm": 2.653731346130371, "learning_rate": 1.999098966204682e-05, "loss": 0.8071, "step": 111 }, { "epoch": 0.33987671882408726, "grad_norm": 2.484330177307129, "learning_rate": 1.9989425643007475e-05, "loss": 0.8135, "step": 112 }, { "epoch": 0.3429113323850166, "grad_norm": 2.3897571563720703, "learning_rate": 1.9987736594069417e-05, "loss": 0.8066, "step": 113 }, { "epoch": 0.34594594594594597, "grad_norm": 2.653904676437378, "learning_rate": 1.998592253637315e-05, "loss": 0.7908, "step": 114 }, { "epoch": 0.3489805595068753, "grad_norm": 2.2755186557769775, "learning_rate": 1.9983983492623832e-05, "loss": 0.7824, "step": 115 }, { "epoch": 0.35201517306780467, "grad_norm": 2.260126829147339, "learning_rate": 1.9981919487090973e-05, "loss": 0.7916, "step": 116 }, { "epoch": 0.355049786628734, "grad_norm": 2.6538352966308594, "learning_rate": 1.9979730545608128e-05, "loss": 0.7927, "step": 117 }, { "epoch": 0.35808440018966337, "grad_norm": 2.2571558952331543, "learning_rate": 1.9977416695572577e-05, "loss": 0.7826, "step": 118 }, { "epoch": 0.3611190137505927, "grad_norm": 2.5154271125793457, "learning_rate": 1.9974977965945e-05, "loss": 0.807, "step": 119 }, { "epoch": 0.36415362731152207, "grad_norm": 2.4908971786499023, "learning_rate": 1.9972414387249074e-05, "loss": 0.7831, "step": 120 }, { "epoch": 0.3671882408724514, "grad_norm": 2.3925859928131104, "learning_rate": 1.996972599157113e-05, "loss": 0.7844, "step": 121 }, { "epoch": 0.37022285443338077, "grad_norm": 2.3527793884277344, "learning_rate": 1.9966912812559733e-05, "loss": 0.7921, "step": 122 }, { "epoch": 0.3732574679943101, "grad_norm": 2.2694365978240967, "learning_rate": 1.9963974885425267e-05, "loss": 0.7816, "step": 123 }, { "epoch": 0.37629208155523947, "grad_norm": 2.436676025390625, "learning_rate": 1.9960912246939486e-05, "loss": 0.7782, "step": 124 }, { "epoch": 0.3793266951161688, "grad_norm": 2.3587653636932373, "learning_rate": 1.9957724935435065e-05, "loss": 0.8024, "step": 125 }, { "epoch": 0.38236130867709817, "grad_norm": 2.3145172595977783, "learning_rate": 1.9954412990805107e-05, "loss": 0.8115, "step": 126 }, { "epoch": 0.3853959222380275, "grad_norm": 2.271946430206299, "learning_rate": 1.995097645450266e-05, "loss": 0.7975, "step": 127 }, { "epoch": 0.38843053579895687, "grad_norm": 2.2533860206604004, "learning_rate": 1.994741536954019e-05, "loss": 0.8187, "step": 128 }, { "epoch": 0.3914651493598862, "grad_norm": 2.5504581928253174, "learning_rate": 1.994372978048903e-05, "loss": 0.7913, "step": 129 }, { "epoch": 0.39449976292081557, "grad_norm": 2.3467888832092285, "learning_rate": 1.993991973347884e-05, "loss": 0.7955, "step": 130 }, { "epoch": 0.3975343764817449, "grad_norm": 2.356933832168579, "learning_rate": 1.9935985276197033e-05, "loss": 0.808, "step": 131 }, { "epoch": 0.40056899004267427, "grad_norm": 2.608546257019043, "learning_rate": 1.9931926457888155e-05, "loss": 0.785, "step": 132 }, { "epoch": 0.4036036036036036, "grad_norm": 2.2730495929718018, "learning_rate": 1.9927743329353295e-05, "loss": 0.79, "step": 133 }, { "epoch": 0.40663821716453297, "grad_norm": 2.2720224857330322, "learning_rate": 1.992343594294943e-05, "loss": 0.8084, "step": 134 }, { "epoch": 0.4096728307254623, "grad_norm": 2.2325122356414795, "learning_rate": 1.9919004352588768e-05, "loss": 0.8008, "step": 135 }, { "epoch": 0.41270744428639167, "grad_norm": 2.4259414672851562, "learning_rate": 1.9914448613738107e-05, "loss": 0.7827, "step": 136 }, { "epoch": 0.415742057847321, "grad_norm": 2.2663819789886475, "learning_rate": 1.9909768783418086e-05, "loss": 0.8059, "step": 137 }, { "epoch": 0.41877667140825037, "grad_norm": 2.910830020904541, "learning_rate": 1.990496492020252e-05, "loss": 0.8159, "step": 138 }, { "epoch": 0.4218112849691797, "grad_norm": 2.4485902786254883, "learning_rate": 1.9900037084217637e-05, "loss": 0.7921, "step": 139 }, { "epoch": 0.42484589853010907, "grad_norm": 2.3199424743652344, "learning_rate": 1.989498533714135e-05, "loss": 0.8006, "step": 140 }, { "epoch": 0.4278805120910384, "grad_norm": 2.529900550842285, "learning_rate": 1.9889809742202454e-05, "loss": 0.7805, "step": 141 }, { "epoch": 0.43091512565196777, "grad_norm": 2.5135438442230225, "learning_rate": 1.988451036417986e-05, "loss": 0.8088, "step": 142 }, { "epoch": 0.4339497392128971, "grad_norm": 2.359200954437256, "learning_rate": 1.9879087269401782e-05, "loss": 0.7963, "step": 143 }, { "epoch": 0.43698435277382647, "grad_norm": 2.307164192199707, "learning_rate": 1.9873540525744888e-05, "loss": 0.79, "step": 144 }, { "epoch": 0.4400189663347558, "grad_norm": 2.341834545135498, "learning_rate": 1.986787020263347e-05, "loss": 0.7955, "step": 145 }, { "epoch": 0.44305357989568517, "grad_norm": 2.409449338912964, "learning_rate": 1.986207637103857e-05, "loss": 0.7761, "step": 146 }, { "epoch": 0.4460881934566145, "grad_norm": 2.3119232654571533, "learning_rate": 1.9856159103477085e-05, "loss": 0.7992, "step": 147 }, { "epoch": 0.44912280701754387, "grad_norm": 2.3376715183258057, "learning_rate": 1.9850118474010873e-05, "loss": 0.784, "step": 148 }, { "epoch": 0.4521574205784732, "grad_norm": 2.387392997741699, "learning_rate": 1.98439545582458e-05, "loss": 0.7806, "step": 149 }, { "epoch": 0.45519203413940257, "grad_norm": 2.2223150730133057, "learning_rate": 1.983766743333084e-05, "loss": 0.7914, "step": 150 }, { "epoch": 0.45519203413940257, "eval_loss": NaN, "eval_runtime": 205.0592, "eval_samples_per_second": 20.277, "eval_steps_per_second": 5.072, "step": 150 }, { "epoch": 0.4582266477003319, "grad_norm": 2.727046012878418, "learning_rate": 1.9831257177957045e-05, "loss": 0.793, "step": 151 }, { "epoch": 0.46126126126126127, "grad_norm": 2.266788959503174, "learning_rate": 1.9824723872356623e-05, "loss": 0.8127, "step": 152 }, { "epoch": 0.4642958748221906, "grad_norm": 2.530904531478882, "learning_rate": 1.9818067598301894e-05, "loss": 0.7973, "step": 153 }, { "epoch": 0.46733048838311997, "grad_norm": 2.3305609226226807, "learning_rate": 1.981128843910428e-05, "loss": 0.7961, "step": 154 }, { "epoch": 0.4703651019440493, "grad_norm": 2.1240079402923584, "learning_rate": 1.9804386479613268e-05, "loss": 0.8093, "step": 155 }, { "epoch": 0.47339971550497867, "grad_norm": 2.625185489654541, "learning_rate": 1.9797361806215335e-05, "loss": 0.7988, "step": 156 }, { "epoch": 0.476434329065908, "grad_norm": 2.4407734870910645, "learning_rate": 1.9790214506832868e-05, "loss": 0.8166, "step": 157 }, { "epoch": 0.47946894262683737, "grad_norm": 2.4013476371765137, "learning_rate": 1.9782944670923075e-05, "loss": 0.7935, "step": 158 }, { "epoch": 0.4825035561877667, "grad_norm": 2.357010841369629, "learning_rate": 1.9775552389476865e-05, "loss": 0.7818, "step": 159 }, { "epoch": 0.48553816974869607, "grad_norm": 2.1862401962280273, "learning_rate": 1.9768037755017687e-05, "loss": 0.7879, "step": 160 }, { "epoch": 0.4885727833096254, "grad_norm": 2.3130927085876465, "learning_rate": 1.97604008616004e-05, "loss": 0.7914, "step": 161 }, { "epoch": 0.49160739687055477, "grad_norm": 2.0661509037017822, "learning_rate": 1.9752641804810083e-05, "loss": 0.8023, "step": 162 }, { "epoch": 0.4946420104314841, "grad_norm": 2.2117955684661865, "learning_rate": 1.9744760681760832e-05, "loss": 0.7972, "step": 163 }, { "epoch": 0.49767662399241347, "grad_norm": 2.603163242340088, "learning_rate": 1.973675759109456e-05, "loss": 0.7913, "step": 164 }, { "epoch": 0.5007112375533428, "grad_norm": 2.2511062622070312, "learning_rate": 1.9728632632979746e-05, "loss": 0.7914, "step": 165 }, { "epoch": 0.5037458511142722, "grad_norm": 2.375213861465454, "learning_rate": 1.9720385909110197e-05, "loss": 0.7928, "step": 166 }, { "epoch": 0.5067804646752015, "grad_norm": 2.5348660945892334, "learning_rate": 1.9712017522703764e-05, "loss": 0.7894, "step": 167 }, { "epoch": 0.5098150782361308, "grad_norm": 2.2094035148620605, "learning_rate": 1.9703527578501052e-05, "loss": 0.7813, "step": 168 }, { "epoch": 0.5128496917970602, "grad_norm": 2.2283380031585693, "learning_rate": 1.9694916182764113e-05, "loss": 0.7877, "step": 169 }, { "epoch": 0.5158843053579896, "grad_norm": 2.189119338989258, "learning_rate": 1.9686183443275118e-05, "loss": 0.7989, "step": 170 }, { "epoch": 0.518918918918919, "grad_norm": 2.2636640071868896, "learning_rate": 1.967732946933499e-05, "loss": 0.8059, "step": 171 }, { "epoch": 0.5219535324798482, "grad_norm": 2.243251085281372, "learning_rate": 1.9668354371762066e-05, "loss": 0.7904, "step": 172 }, { "epoch": 0.5249881460407776, "grad_norm": 2.215536117553711, "learning_rate": 1.9659258262890683e-05, "loss": 0.7912, "step": 173 }, { "epoch": 0.528022759601707, "grad_norm": 2.2998199462890625, "learning_rate": 1.9650041256569792e-05, "loss": 0.797, "step": 174 }, { "epoch": 0.5310573731626363, "grad_norm": 2.3586950302124023, "learning_rate": 1.9640703468161508e-05, "loss": 0.7907, "step": 175 }, { "epoch": 0.5340919867235657, "grad_norm": 2.257404088973999, "learning_rate": 1.96312450145397e-05, "loss": 0.7977, "step": 176 }, { "epoch": 0.537126600284495, "grad_norm": 2.4808526039123535, "learning_rate": 1.9621666014088495e-05, "loss": 0.7929, "step": 177 }, { "epoch": 0.5401612138454244, "grad_norm": 2.208704710006714, "learning_rate": 1.9611966586700825e-05, "loss": 0.7975, "step": 178 }, { "epoch": 0.5431958274063537, "grad_norm": 2.4615161418914795, "learning_rate": 1.9602146853776894e-05, "loss": 0.7991, "step": 179 }, { "epoch": 0.5462304409672831, "grad_norm": 2.4766757488250732, "learning_rate": 1.9592206938222703e-05, "loss": 0.7911, "step": 180 }, { "epoch": 0.5492650545282124, "grad_norm": 2.2009432315826416, "learning_rate": 1.9582146964448457e-05, "loss": 0.788, "step": 181 }, { "epoch": 0.5522996680891418, "grad_norm": 2.258129358291626, "learning_rate": 1.9571967058367067e-05, "loss": 0.7893, "step": 182 }, { "epoch": 0.5553342816500711, "grad_norm": 2.1914985179901123, "learning_rate": 1.956166734739251e-05, "loss": 0.8057, "step": 183 }, { "epoch": 0.5583688952110005, "grad_norm": 2.2815279960632324, "learning_rate": 1.9551247960438298e-05, "loss": 0.7823, "step": 184 }, { "epoch": 0.5614035087719298, "grad_norm": 2.2393579483032227, "learning_rate": 1.954070902791582e-05, "loss": 0.7899, "step": 185 }, { "epoch": 0.5644381223328592, "grad_norm": 2.56640625, "learning_rate": 1.953005068173272e-05, "loss": 0.7731, "step": 186 }, { "epoch": 0.5674727358937885, "grad_norm": 2.3896234035491943, "learning_rate": 1.9519273055291266e-05, "loss": 0.7936, "step": 187 }, { "epoch": 0.5705073494547179, "grad_norm": 2.21494722366333, "learning_rate": 1.9508376283486653e-05, "loss": 0.8121, "step": 188 }, { "epoch": 0.5735419630156472, "grad_norm": 2.400538444519043, "learning_rate": 1.949736050270532e-05, "loss": 0.7938, "step": 189 }, { "epoch": 0.5765765765765766, "grad_norm": 2.1337621212005615, "learning_rate": 1.9486225850823265e-05, "loss": 0.8049, "step": 190 }, { "epoch": 0.579611190137506, "grad_norm": 2.314168930053711, "learning_rate": 1.9474972467204298e-05, "loss": 0.8109, "step": 191 }, { "epoch": 0.5826458036984353, "grad_norm": 2.2364132404327393, "learning_rate": 1.9463600492698297e-05, "loss": 0.78, "step": 192 }, { "epoch": 0.5856804172593646, "grad_norm": 2.204206705093384, "learning_rate": 1.945211006963945e-05, "loss": 0.8187, "step": 193 }, { "epoch": 0.588715030820294, "grad_norm": 2.1865625381469727, "learning_rate": 1.9440501341844484e-05, "loss": 0.7844, "step": 194 }, { "epoch": 0.5917496443812233, "grad_norm": 2.143092155456543, "learning_rate": 1.9428774454610845e-05, "loss": 0.7967, "step": 195 }, { "epoch": 0.5947842579421527, "grad_norm": 2.3440985679626465, "learning_rate": 1.9416929554714887e-05, "loss": 0.8096, "step": 196 }, { "epoch": 0.597818871503082, "grad_norm": 2.43977689743042, "learning_rate": 1.9404966790410047e-05, "loss": 0.795, "step": 197 }, { "epoch": 0.6008534850640114, "grad_norm": 2.2740390300750732, "learning_rate": 1.9392886311424975e-05, "loss": 0.7937, "step": 198 }, { "epoch": 0.6038880986249408, "grad_norm": 2.1775126457214355, "learning_rate": 1.938068826896166e-05, "loss": 0.8019, "step": 199 }, { "epoch": 0.6069227121858701, "grad_norm": 2.1858487129211426, "learning_rate": 1.9368372815693547e-05, "loss": 0.7962, "step": 200 }, { "epoch": 0.6069227121858701, "eval_loss": NaN, "eval_runtime": 204.6421, "eval_samples_per_second": 20.318, "eval_steps_per_second": 5.082, "step": 200 }, { "epoch": 0.6099573257467994, "grad_norm": 2.2958755493164062, "learning_rate": 1.9355940105763622e-05, "loss": 0.8003, "step": 201 }, { "epoch": 0.6129919393077288, "grad_norm": 2.1330931186676025, "learning_rate": 1.934339029478248e-05, "loss": 0.7803, "step": 202 }, { "epoch": 0.6160265528686582, "grad_norm": 2.3881402015686035, "learning_rate": 1.9330723539826373e-05, "loss": 0.8044, "step": 203 }, { "epoch": 0.6190611664295875, "grad_norm": 2.2404513359069824, "learning_rate": 1.9317939999435262e-05, "loss": 0.8097, "step": 204 }, { "epoch": 0.6220957799905168, "grad_norm": 2.194645404815674, "learning_rate": 1.930503983361081e-05, "loss": 0.8178, "step": 205 }, { "epoch": 0.6251303935514462, "grad_norm": 2.509723424911499, "learning_rate": 1.92920232038144e-05, "loss": 0.7936, "step": 206 }, { "epoch": 0.6281650071123756, "grad_norm": 2.2457869052886963, "learning_rate": 1.9278890272965097e-05, "loss": 0.8162, "step": 207 }, { "epoch": 0.6311996206733049, "grad_norm": 2.2989683151245117, "learning_rate": 1.9265641205437612e-05, "loss": 0.8012, "step": 208 }, { "epoch": 0.6342342342342342, "grad_norm": 2.3188092708587646, "learning_rate": 1.925227616706026e-05, "loss": 0.8094, "step": 209 }, { "epoch": 0.6372688477951636, "grad_norm": 2.2978076934814453, "learning_rate": 1.9238795325112867e-05, "loss": 0.7811, "step": 210 }, { "epoch": 0.640303461356093, "grad_norm": 2.331542730331421, "learning_rate": 1.9225198848324687e-05, "loss": 0.7946, "step": 211 }, { "epoch": 0.6433380749170223, "grad_norm": 2.1989738941192627, "learning_rate": 1.921148690687228e-05, "loss": 0.7983, "step": 212 }, { "epoch": 0.6463726884779516, "grad_norm": 2.4261419773101807, "learning_rate": 1.9197659672377388e-05, "loss": 0.8134, "step": 213 }, { "epoch": 0.649407302038881, "grad_norm": 2.3790082931518555, "learning_rate": 1.918371731790479e-05, "loss": 0.79, "step": 214 }, { "epoch": 0.6524419155998104, "grad_norm": 2.2966949939727783, "learning_rate": 1.9169660017960135e-05, "loss": 0.8029, "step": 215 }, { "epoch": 0.6554765291607397, "grad_norm": 2.5911426544189453, "learning_rate": 1.915548794848775e-05, "loss": 0.8118, "step": 216 }, { "epoch": 0.658511142721669, "grad_norm": 2.3282856941223145, "learning_rate": 1.9141201286868435e-05, "loss": 0.8092, "step": 217 }, { "epoch": 0.6615457562825984, "grad_norm": 2.415398359298706, "learning_rate": 1.9126800211917277e-05, "loss": 0.8156, "step": 218 }, { "epoch": 0.6645803698435278, "grad_norm": 2.2823410034179688, "learning_rate": 1.911228490388136e-05, "loss": 0.8004, "step": 219 }, { "epoch": 0.6676149834044571, "grad_norm": 2.2104527950286865, "learning_rate": 1.9097655544437544e-05, "loss": 0.8023, "step": 220 }, { "epoch": 0.6706495969653864, "grad_norm": 2.3354063034057617, "learning_rate": 1.908291231669019e-05, "loss": 0.8117, "step": 221 }, { "epoch": 0.6736842105263158, "grad_norm": 2.230656147003174, "learning_rate": 1.906805540516885e-05, "loss": 0.7797, "step": 222 }, { "epoch": 0.6767188240872452, "grad_norm": 2.464111328125, "learning_rate": 1.905308499582597e-05, "loss": 0.7929, "step": 223 }, { "epoch": 0.6797534376481745, "grad_norm": 2.188788652420044, "learning_rate": 1.903800127603456e-05, "loss": 0.7971, "step": 224 }, { "epoch": 0.6827880512091038, "grad_norm": 2.202427387237549, "learning_rate": 1.9022804434585854e-05, "loss": 0.8026, "step": 225 }, { "epoch": 0.6858226647700332, "grad_norm": 2.2621190547943115, "learning_rate": 1.9007494661686937e-05, "loss": 0.8112, "step": 226 }, { "epoch": 0.6888572783309626, "grad_norm": 2.333603620529175, "learning_rate": 1.8992072148958368e-05, "loss": 0.7937, "step": 227 }, { "epoch": 0.6918918918918919, "grad_norm": 2.626451253890991, "learning_rate": 1.8976537089431793e-05, "loss": 0.8005, "step": 228 }, { "epoch": 0.6949265054528212, "grad_norm": 2.4107227325439453, "learning_rate": 1.8960889677547506e-05, "loss": 0.7813, "step": 229 }, { "epoch": 0.6979611190137506, "grad_norm": 2.7147607803344727, "learning_rate": 1.8945130109152035e-05, "loss": 0.8036, "step": 230 }, { "epoch": 0.70099573257468, "grad_norm": 2.281703233718872, "learning_rate": 1.8929258581495688e-05, "loss": 0.7946, "step": 231 }, { "epoch": 0.7040303461356093, "grad_norm": 3.016942262649536, "learning_rate": 1.891327529323007e-05, "loss": 0.7786, "step": 232 }, { "epoch": 0.7070649596965386, "grad_norm": 2.5729317665100098, "learning_rate": 1.8897180444405615e-05, "loss": 0.8141, "step": 233 }, { "epoch": 0.710099573257468, "grad_norm": 2.75722336769104, "learning_rate": 1.888097423646907e-05, "loss": 0.8079, "step": 234 }, { "epoch": 0.7131341868183974, "grad_norm": 2.843980073928833, "learning_rate": 1.8864656872260985e-05, "loss": 0.795, "step": 235 }, { "epoch": 0.7161688003793267, "grad_norm": 3.3999879360198975, "learning_rate": 1.884822855601316e-05, "loss": 0.8086, "step": 236 }, { "epoch": 0.719203413940256, "grad_norm": 3.1997487545013428, "learning_rate": 1.8831689493346095e-05, "loss": 0.7919, "step": 237 }, { "epoch": 0.7222380275011854, "grad_norm": 2.4851367473602295, "learning_rate": 1.881503989126642e-05, "loss": 0.7914, "step": 238 }, { "epoch": 0.7252726410621148, "grad_norm": 2.738428831100464, "learning_rate": 1.8798279958164295e-05, "loss": 0.7929, "step": 239 }, { "epoch": 0.7283072546230441, "grad_norm": 2.1467783451080322, "learning_rate": 1.8781409903810823e-05, "loss": 0.815, "step": 240 }, { "epoch": 0.7313418681839734, "grad_norm": 16.773521423339844, "learning_rate": 1.8764429939355394e-05, "loss": 0.8014, "step": 241 }, { "epoch": 0.7343764817449028, "grad_norm": 3.171633720397949, "learning_rate": 1.874734027732306e-05, "loss": 0.805, "step": 242 }, { "epoch": 0.7374110953058322, "grad_norm": 2.249514102935791, "learning_rate": 1.8730141131611882e-05, "loss": 0.8109, "step": 243 }, { "epoch": 0.7404457088667615, "grad_norm": 2.816488742828369, "learning_rate": 1.8712832717490238e-05, "loss": 0.8044, "step": 244 }, { "epoch": 0.7434803224276908, "grad_norm": 2.558295488357544, "learning_rate": 1.8695415251594123e-05, "loss": 0.8269, "step": 245 }, { "epoch": 0.7465149359886202, "grad_norm": 2.5777034759521484, "learning_rate": 1.8677888951924473e-05, "loss": 0.7971, "step": 246 }, { "epoch": 0.7495495495495496, "grad_norm": 2.394287586212158, "learning_rate": 1.866025403784439e-05, "loss": 0.8092, "step": 247 }, { "epoch": 0.7525841631104789, "grad_norm": 2.2798614501953125, "learning_rate": 1.864251073007642e-05, "loss": 0.7964, "step": 248 }, { "epoch": 0.7556187766714082, "grad_norm": 2.3587262630462646, "learning_rate": 1.8624659250699807e-05, "loss": 0.7928, "step": 249 }, { "epoch": 0.7586533902323376, "grad_norm": 2.189763307571411, "learning_rate": 1.8606699823147675e-05, "loss": 0.7812, "step": 250 }, { "epoch": 0.7586533902323376, "eval_loss": NaN, "eval_runtime": 204.0596, "eval_samples_per_second": 20.376, "eval_steps_per_second": 5.097, "step": 250 }, { "epoch": 0.761688003793267, "grad_norm": 2.433803081512451, "learning_rate": 1.8588632672204264e-05, "loss": 0.8111, "step": 251 }, { "epoch": 0.7647226173541963, "grad_norm": 2.28267765045166, "learning_rate": 1.8570458024002094e-05, "loss": 0.8001, "step": 252 }, { "epoch": 0.7677572309151256, "grad_norm": 2.3339545726776123, "learning_rate": 1.8552176106019156e-05, "loss": 0.8158, "step": 253 }, { "epoch": 0.770791844476055, "grad_norm": 2.284759998321533, "learning_rate": 1.8533787147076046e-05, "loss": 0.7852, "step": 254 }, { "epoch": 0.7738264580369844, "grad_norm": 2.35969614982605, "learning_rate": 1.8515291377333114e-05, "loss": 0.7909, "step": 255 }, { "epoch": 0.7768610715979137, "grad_norm": 2.4525341987609863, "learning_rate": 1.8496689028287572e-05, "loss": 0.817, "step": 256 }, { "epoch": 0.779895685158843, "grad_norm": 2.7126755714416504, "learning_rate": 1.847798033277061e-05, "loss": 0.797, "step": 257 }, { "epoch": 0.7829302987197724, "grad_norm": 2.2292561531066895, "learning_rate": 1.8459165524944463e-05, "loss": 0.8044, "step": 258 }, { "epoch": 0.7859649122807018, "grad_norm": 2.447347640991211, "learning_rate": 1.8440244840299507e-05, "loss": 0.7979, "step": 259 }, { "epoch": 0.7889995258416311, "grad_norm": 2.2336087226867676, "learning_rate": 1.842121851565128e-05, "loss": 0.8036, "step": 260 }, { "epoch": 0.7920341394025604, "grad_norm": 2.3720502853393555, "learning_rate": 1.8402086789137547e-05, "loss": 0.7979, "step": 261 }, { "epoch": 0.7950687529634898, "grad_norm": 2.1780807971954346, "learning_rate": 1.8382849900215297e-05, "loss": 0.7876, "step": 262 }, { "epoch": 0.7981033665244192, "grad_norm": 2.3325858116149902, "learning_rate": 1.8363508089657763e-05, "loss": 0.7997, "step": 263 }, { "epoch": 0.8011379800853485, "grad_norm": 2.3341164588928223, "learning_rate": 1.8344061599551397e-05, "loss": 0.7844, "step": 264 }, { "epoch": 0.8041725936462779, "grad_norm": 2.4678280353546143, "learning_rate": 1.8324510673292844e-05, "loss": 0.7946, "step": 265 }, { "epoch": 0.8072072072072072, "grad_norm": 2.424893856048584, "learning_rate": 1.8304855555585893e-05, "loss": 0.7916, "step": 266 }, { "epoch": 0.8102418207681366, "grad_norm": 2.5433976650238037, "learning_rate": 1.8285096492438424e-05, "loss": 0.7983, "step": 267 }, { "epoch": 0.8132764343290659, "grad_norm": 2.499178647994995, "learning_rate": 1.826523373115931e-05, "loss": 0.7944, "step": 268 }, { "epoch": 0.8163110478899953, "grad_norm": 2.2453994750976562, "learning_rate": 1.8245267520355348e-05, "loss": 0.8148, "step": 269 }, { "epoch": 0.8193456614509246, "grad_norm": 2.350146770477295, "learning_rate": 1.8225198109928116e-05, "loss": 0.7986, "step": 270 }, { "epoch": 0.822380275011854, "grad_norm": 2.2048559188842773, "learning_rate": 1.8205025751070878e-05, "loss": 0.8093, "step": 271 }, { "epoch": 0.8254148885727833, "grad_norm": 2.2730185985565186, "learning_rate": 1.8184750696265408e-05, "loss": 0.7787, "step": 272 }, { "epoch": 0.8284495021337127, "grad_norm": 2.423301935195923, "learning_rate": 1.8164373199278858e-05, "loss": 0.823, "step": 273 }, { "epoch": 0.831484115694642, "grad_norm": 2.309649705886841, "learning_rate": 1.8143893515160565e-05, "loss": 0.7901, "step": 274 }, { "epoch": 0.8345187292555714, "grad_norm": 2.249284267425537, "learning_rate": 1.812331190023886e-05, "loss": 0.8095, "step": 275 }, { "epoch": 0.8375533428165007, "grad_norm": 2.2063703536987305, "learning_rate": 1.8102628612117868e-05, "loss": 0.8008, "step": 276 }, { "epoch": 0.8405879563774301, "grad_norm": 2.2518839836120605, "learning_rate": 1.8081843909674277e-05, "loss": 0.8051, "step": 277 }, { "epoch": 0.8436225699383594, "grad_norm": 2.226356267929077, "learning_rate": 1.8060958053054095e-05, "loss": 0.8036, "step": 278 }, { "epoch": 0.8466571834992888, "grad_norm": 2.186485767364502, "learning_rate": 1.8039971303669407e-05, "loss": 0.8025, "step": 279 }, { "epoch": 0.8496917970602181, "grad_norm": 2.4235646724700928, "learning_rate": 1.8018883924195085e-05, "loss": 0.7799, "step": 280 }, { "epoch": 0.8527264106211475, "grad_norm": 2.2470943927764893, "learning_rate": 1.799769617856552e-05, "loss": 0.8025, "step": 281 }, { "epoch": 0.8557610241820768, "grad_norm": 2.2514889240264893, "learning_rate": 1.79764083319713e-05, "loss": 0.7978, "step": 282 }, { "epoch": 0.8587956377430062, "grad_norm": 2.220952033996582, "learning_rate": 1.79550206508559e-05, "loss": 0.7978, "step": 283 }, { "epoch": 0.8618302513039355, "grad_norm": 2.4699270725250244, "learning_rate": 1.7933533402912354e-05, "loss": 0.7835, "step": 284 }, { "epoch": 0.8648648648648649, "grad_norm": 2.3011207580566406, "learning_rate": 1.7911946857079886e-05, "loss": 0.8008, "step": 285 }, { "epoch": 0.8678994784257942, "grad_norm": 2.2239327430725098, "learning_rate": 1.7890261283540563e-05, "loss": 0.7904, "step": 286 }, { "epoch": 0.8709340919867236, "grad_norm": 2.077845573425293, "learning_rate": 1.78684769537159e-05, "loss": 0.7962, "step": 287 }, { "epoch": 0.8739687055476529, "grad_norm": 2.2492687702178955, "learning_rate": 1.7846594140263475e-05, "loss": 0.8076, "step": 288 }, { "epoch": 0.8770033191085823, "grad_norm": 2.1000773906707764, "learning_rate": 1.78246131170735e-05, "loss": 0.8107, "step": 289 }, { "epoch": 0.8800379326695116, "grad_norm": 2.2440242767333984, "learning_rate": 1.7802534159265407e-05, "loss": 0.784, "step": 290 }, { "epoch": 0.883072546230441, "grad_norm": 2.1370065212249756, "learning_rate": 1.7780357543184396e-05, "loss": 0.7926, "step": 291 }, { "epoch": 0.8861071597913703, "grad_norm": 2.2182703018188477, "learning_rate": 1.775808354639799e-05, "loss": 0.8079, "step": 292 }, { "epoch": 0.8891417733522997, "grad_norm": 2.236370325088501, "learning_rate": 1.773571244769254e-05, "loss": 0.7851, "step": 293 }, { "epoch": 0.892176386913229, "grad_norm": 2.216042995452881, "learning_rate": 1.771324452706975e-05, "loss": 0.798, "step": 294 }, { "epoch": 0.8952110004741584, "grad_norm": 2.2804715633392334, "learning_rate": 1.769068006574317e-05, "loss": 0.7916, "step": 295 }, { "epoch": 0.8982456140350877, "grad_norm": 2.188271999359131, "learning_rate": 1.7668019346134674e-05, "loss": 0.7993, "step": 296 }, { "epoch": 0.9012802275960171, "grad_norm": 2.372596025466919, "learning_rate": 1.7645262651870926e-05, "loss": 0.816, "step": 297 }, { "epoch": 0.9043148411569464, "grad_norm": 2.174302339553833, "learning_rate": 1.7622410267779834e-05, "loss": 0.8247, "step": 298 }, { "epoch": 0.9073494547178758, "grad_norm": 2.0944302082061768, "learning_rate": 1.7599462479886976e-05, "loss": 0.7979, "step": 299 }, { "epoch": 0.9103840682788051, "grad_norm": 2.118502140045166, "learning_rate": 1.7576419575412028e-05, "loss": 0.8007, "step": 300 }, { "epoch": 0.9103840682788051, "eval_loss": NaN, "eval_runtime": 203.6703, "eval_samples_per_second": 20.415, "eval_steps_per_second": 5.106, "step": 300 }, { "epoch": 0.9134186818397345, "grad_norm": 2.2053334712982178, "learning_rate": 1.755328184276517e-05, "loss": 0.7913, "step": 301 }, { "epoch": 0.9164532954006638, "grad_norm": 2.243788957595825, "learning_rate": 1.7530049571543464e-05, "loss": 0.801, "step": 302 }, { "epoch": 0.9194879089615932, "grad_norm": 2.363306760787964, "learning_rate": 1.7506723052527243e-05, "loss": 0.8278, "step": 303 }, { "epoch": 0.9225225225225225, "grad_norm": 2.2566967010498047, "learning_rate": 1.7483302577676475e-05, "loss": 0.7929, "step": 304 }, { "epoch": 0.9255571360834519, "grad_norm": 2.1566100120544434, "learning_rate": 1.7459788440127083e-05, "loss": 0.7953, "step": 305 }, { "epoch": 0.9285917496443812, "grad_norm": 2.279130697250366, "learning_rate": 1.7436180934187307e-05, "loss": 0.8125, "step": 306 }, { "epoch": 0.9316263632053106, "grad_norm": 2.172891616821289, "learning_rate": 1.7412480355334006e-05, "loss": 0.8007, "step": 307 }, { "epoch": 0.9346609767662399, "grad_norm": 2.2873098850250244, "learning_rate": 1.738868700020895e-05, "loss": 0.779, "step": 308 }, { "epoch": 0.9376955903271693, "grad_norm": 2.3155357837677, "learning_rate": 1.7364801166615124e-05, "loss": 0.8025, "step": 309 }, { "epoch": 0.9407302038880986, "grad_norm": 2.20151686668396, "learning_rate": 1.7340823153513003e-05, "loss": 0.8173, "step": 310 }, { "epoch": 0.943764817449028, "grad_norm": 2.2033958435058594, "learning_rate": 1.7316753261016782e-05, "loss": 0.8095, "step": 311 }, { "epoch": 0.9467994310099573, "grad_norm": 2.2805070877075195, "learning_rate": 1.7292591790390668e-05, "loss": 0.8139, "step": 312 }, { "epoch": 0.9498340445708867, "grad_norm": 2.335238456726074, "learning_rate": 1.7268339044045044e-05, "loss": 0.7898, "step": 313 }, { "epoch": 0.952868658131816, "grad_norm": 2.1889333724975586, "learning_rate": 1.7243995325532755e-05, "loss": 0.8002, "step": 314 }, { "epoch": 0.9559032716927454, "grad_norm": 2.5007987022399902, "learning_rate": 1.7219560939545246e-05, "loss": 0.7892, "step": 315 }, { "epoch": 0.9589378852536747, "grad_norm": 2.1342813968658447, "learning_rate": 1.7195036191908798e-05, "loss": 0.8028, "step": 316 }, { "epoch": 0.9619724988146041, "grad_norm": 2.199307918548584, "learning_rate": 1.7170421389580666e-05, "loss": 0.793, "step": 317 }, { "epoch": 0.9650071123755334, "grad_norm": 2.1955904960632324, "learning_rate": 1.7145716840645253e-05, "loss": 0.8085, "step": 318 }, { "epoch": 0.9680417259364628, "grad_norm": 2.41741943359375, "learning_rate": 1.712092285431026e-05, "loss": 0.7964, "step": 319 }, { "epoch": 0.9710763394973921, "grad_norm": 2.3595402240753174, "learning_rate": 1.7096039740902782e-05, "loss": 0.7999, "step": 320 }, { "epoch": 0.9741109530583215, "grad_norm": 2.15049147605896, "learning_rate": 1.7071067811865477e-05, "loss": 0.7813, "step": 321 }, { "epoch": 0.9771455666192508, "grad_norm": 2.240618944168091, "learning_rate": 1.7046007379752624e-05, "loss": 0.8038, "step": 322 }, { "epoch": 0.9801801801801802, "grad_norm": 2.418973922729492, "learning_rate": 1.702085875822623e-05, "loss": 0.7672, "step": 323 }, { "epoch": 0.9832147937411095, "grad_norm": 2.2924294471740723, "learning_rate": 1.6995622262052093e-05, "loss": 0.8013, "step": 324 }, { "epoch": 0.9862494073020389, "grad_norm": 2.2135136127471924, "learning_rate": 1.6970298207095887e-05, "loss": 0.8112, "step": 325 }, { "epoch": 0.9892840208629682, "grad_norm": 2.2720751762390137, "learning_rate": 1.6944886910319173e-05, "loss": 0.7896, "step": 326 }, { "epoch": 0.9923186344238976, "grad_norm": 2.172165632247925, "learning_rate": 1.6919388689775463e-05, "loss": 0.8084, "step": 327 }, { "epoch": 0.9953532479848269, "grad_norm": 2.1608591079711914, "learning_rate": 1.6893803864606224e-05, "loss": 0.7914, "step": 328 }, { "epoch": 0.9983878615457563, "grad_norm": 2.153231143951416, "learning_rate": 1.6868132755036875e-05, "loss": 0.803, "step": 329 }, { "epoch": 1.0014224751066856, "grad_norm": 2.242035150527954, "learning_rate": 1.6842375682372803e-05, "loss": 0.7964, "step": 330 }, { "epoch": 1.004457088667615, "grad_norm": 2.490118980407715, "learning_rate": 1.681653296899533e-05, "loss": 0.7645, "step": 331 }, { "epoch": 1.0074917022285443, "grad_norm": 2.258561372756958, "learning_rate": 1.6790604938357664e-05, "loss": 0.7579, "step": 332 }, { "epoch": 1.0105263157894737, "grad_norm": 2.3750267028808594, "learning_rate": 1.676459191498087e-05, "loss": 0.7849, "step": 333 }, { "epoch": 1.013560929350403, "grad_norm": 2.5088164806365967, "learning_rate": 1.6738494224449802e-05, "loss": 0.7814, "step": 334 }, { "epoch": 1.0165955429113325, "grad_norm": 2.4220824241638184, "learning_rate": 1.6712312193409032e-05, "loss": 0.7774, "step": 335 }, { "epoch": 1.0196301564722616, "grad_norm": 2.2030773162841797, "learning_rate": 1.6686046149558736e-05, "loss": 0.7495, "step": 336 }, { "epoch": 1.022664770033191, "grad_norm": 2.336583375930786, "learning_rate": 1.6659696421650645e-05, "loss": 0.7508, "step": 337 }, { "epoch": 1.0256993835941204, "grad_norm": 2.3112359046936035, "learning_rate": 1.6633263339483867e-05, "loss": 0.7525, "step": 338 }, { "epoch": 1.0287339971550498, "grad_norm": 2.165022611618042, "learning_rate": 1.6606747233900816e-05, "loss": 0.787, "step": 339 }, { "epoch": 1.0317686107159791, "grad_norm": 2.3545596599578857, "learning_rate": 1.658014843678303e-05, "loss": 0.7665, "step": 340 }, { "epoch": 1.0348032242769085, "grad_norm": 2.231351137161255, "learning_rate": 1.655346728104704e-05, "loss": 0.7726, "step": 341 }, { "epoch": 1.037837837837838, "grad_norm": 2.292428970336914, "learning_rate": 1.652670410064019e-05, "loss": 0.7722, "step": 342 }, { "epoch": 1.0408724513987673, "grad_norm": 2.2555713653564453, "learning_rate": 1.6499859230536468e-05, "loss": 0.755, "step": 343 }, { "epoch": 1.0439070649596967, "grad_norm": 2.5129449367523193, "learning_rate": 1.647293300673231e-05, "loss": 0.7736, "step": 344 }, { "epoch": 1.0469416785206258, "grad_norm": 2.269122362136841, "learning_rate": 1.6445925766242392e-05, "loss": 0.7898, "step": 345 }, { "epoch": 1.0499762920815552, "grad_norm": 2.21991229057312, "learning_rate": 1.641883784709541e-05, "loss": 0.7767, "step": 346 }, { "epoch": 1.0530109056424846, "grad_norm": 2.324253797531128, "learning_rate": 1.639166958832985e-05, "loss": 0.7728, "step": 347 }, { "epoch": 1.056045519203414, "grad_norm": 2.3205628395080566, "learning_rate": 1.6364421329989758e-05, "loss": 0.7845, "step": 348 }, { "epoch": 1.0590801327643433, "grad_norm": 2.361678123474121, "learning_rate": 1.6337093413120463e-05, "loss": 0.7455, "step": 349 }, { "epoch": 1.0621147463252727, "grad_norm": 2.3375606536865234, "learning_rate": 1.6309686179764317e-05, "loss": 0.7754, "step": 350 }, { "epoch": 1.0621147463252727, "eval_loss": NaN, "eval_runtime": 203.6306, "eval_samples_per_second": 20.419, "eval_steps_per_second": 5.107, "step": 350 }, { "epoch": 1.065149359886202, "grad_norm": 2.3522422313690186, "learning_rate": 1.6282199972956425e-05, "loss": 0.7759, "step": 351 }, { "epoch": 1.0681839734471312, "grad_norm": 2.4227213859558105, "learning_rate": 1.6254635136720328e-05, "loss": 0.7772, "step": 352 }, { "epoch": 1.0712185870080606, "grad_norm": 2.235722303390503, "learning_rate": 1.6226992016063726e-05, "loss": 0.7694, "step": 353 }, { "epoch": 1.07425320056899, "grad_norm": 2.2417314052581787, "learning_rate": 1.6199270956974128e-05, "loss": 0.7628, "step": 354 }, { "epoch": 1.0772878141299194, "grad_norm": 2.348954200744629, "learning_rate": 1.6171472306414554e-05, "loss": 0.7656, "step": 355 }, { "epoch": 1.0803224276908487, "grad_norm": 2.346963882446289, "learning_rate": 1.614359641231916e-05, "loss": 0.7839, "step": 356 }, { "epoch": 1.0833570412517781, "grad_norm": 2.2969138622283936, "learning_rate": 1.6115643623588915e-05, "loss": 0.7728, "step": 357 }, { "epoch": 1.0863916548127075, "grad_norm": 2.3338327407836914, "learning_rate": 1.608761429008721e-05, "loss": 0.7902, "step": 358 }, { "epoch": 1.0894262683736369, "grad_norm": 2.2462401390075684, "learning_rate": 1.6059508762635482e-05, "loss": 0.7906, "step": 359 }, { "epoch": 1.0924608819345663, "grad_norm": 2.2091758251190186, "learning_rate": 1.6031327393008848e-05, "loss": 0.7587, "step": 360 }, { "epoch": 1.0954954954954954, "grad_norm": 2.2392489910125732, "learning_rate": 1.6003070533931657e-05, "loss": 0.7598, "step": 361 }, { "epoch": 1.0985301090564248, "grad_norm": 2.193833112716675, "learning_rate": 1.5974738539073125e-05, "loss": 0.7622, "step": 362 }, { "epoch": 1.1015647226173542, "grad_norm": 2.187610149383545, "learning_rate": 1.594633176304287e-05, "loss": 0.7796, "step": 363 }, { "epoch": 1.1045993361782835, "grad_norm": 2.2753069400787354, "learning_rate": 1.5917850561386487e-05, "loss": 0.7783, "step": 364 }, { "epoch": 1.107633949739213, "grad_norm": 2.2835614681243896, "learning_rate": 1.588929529058111e-05, "loss": 0.7801, "step": 365 }, { "epoch": 1.1106685633001423, "grad_norm": 2.46498441696167, "learning_rate": 1.5860666308030933e-05, "loss": 0.7683, "step": 366 }, { "epoch": 1.1137031768610717, "grad_norm": 2.361351490020752, "learning_rate": 1.5831963972062734e-05, "loss": 0.783, "step": 367 }, { "epoch": 1.1167377904220008, "grad_norm": 2.2396347522735596, "learning_rate": 1.5803188641921417e-05, "loss": 0.7563, "step": 368 }, { "epoch": 1.1197724039829302, "grad_norm": 2.2810609340667725, "learning_rate": 1.5774340677765483e-05, "loss": 0.7865, "step": 369 }, { "epoch": 1.1228070175438596, "grad_norm": 2.147937297821045, "learning_rate": 1.5745420440662543e-05, "loss": 0.7684, "step": 370 }, { "epoch": 1.125841631104789, "grad_norm": 2.355337142944336, "learning_rate": 1.5716428292584788e-05, "loss": 0.7693, "step": 371 }, { "epoch": 1.1288762446657183, "grad_norm": 2.4234957695007324, "learning_rate": 1.568736459640447e-05, "loss": 0.7567, "step": 372 }, { "epoch": 1.1319108582266477, "grad_norm": 2.2321126461029053, "learning_rate": 1.5658229715889345e-05, "loss": 0.7984, "step": 373 }, { "epoch": 1.134945471787577, "grad_norm": 2.2270772457122803, "learning_rate": 1.5629024015698137e-05, "loss": 0.7868, "step": 374 }, { "epoch": 1.1379800853485065, "grad_norm": 2.4906022548675537, "learning_rate": 1.5599747861375957e-05, "loss": 0.7761, "step": 375 }, { "epoch": 1.1410146989094359, "grad_norm": 2.4099533557891846, "learning_rate": 1.5570401619349737e-05, "loss": 0.7727, "step": 376 }, { "epoch": 1.144049312470365, "grad_norm": 2.167451858520508, "learning_rate": 1.5540985656923648e-05, "loss": 0.7449, "step": 377 }, { "epoch": 1.1470839260312944, "grad_norm": 2.4137990474700928, "learning_rate": 1.551150034227449e-05, "loss": 0.7836, "step": 378 }, { "epoch": 1.1501185395922238, "grad_norm": 2.0170676708221436, "learning_rate": 1.54819460444471e-05, "loss": 0.771, "step": 379 }, { "epoch": 1.1531531531531531, "grad_norm": 2.3556909561157227, "learning_rate": 1.5452323133349712e-05, "loss": 0.769, "step": 380 }, { "epoch": 1.1561877667140825, "grad_norm": 2.1490001678466797, "learning_rate": 1.5422631979749354e-05, "loss": 0.7467, "step": 381 }, { "epoch": 1.159222380275012, "grad_norm": 2.240185260772705, "learning_rate": 1.5392872955267176e-05, "loss": 0.7677, "step": 382 }, { "epoch": 1.1622569938359413, "grad_norm": 2.1402430534362793, "learning_rate": 1.5363046432373824e-05, "loss": 0.7706, "step": 383 }, { "epoch": 1.1652916073968704, "grad_norm": 2.043536424636841, "learning_rate": 1.5333152784384777e-05, "loss": 0.7679, "step": 384 }, { "epoch": 1.1683262209577998, "grad_norm": 2.146432399749756, "learning_rate": 1.5303192385455652e-05, "loss": 0.7746, "step": 385 }, { "epoch": 1.1713608345187292, "grad_norm": 2.216525077819824, "learning_rate": 1.5273165610577543e-05, "loss": 0.7735, "step": 386 }, { "epoch": 1.1743954480796586, "grad_norm": 2.134411334991455, "learning_rate": 1.5243072835572319e-05, "loss": 0.771, "step": 387 }, { "epoch": 1.177430061640588, "grad_norm": 2.210275173187256, "learning_rate": 1.5212914437087921e-05, "loss": 0.7665, "step": 388 }, { "epoch": 1.1804646752015173, "grad_norm": 2.2126176357269287, "learning_rate": 1.5182690792593659e-05, "loss": 0.7658, "step": 389 }, { "epoch": 1.1834992887624467, "grad_norm": 1.9885146617889404, "learning_rate": 1.5152402280375454e-05, "loss": 0.7509, "step": 390 }, { "epoch": 1.186533902323376, "grad_norm": 2.2027952671051025, "learning_rate": 1.5122049279531143e-05, "loss": 0.7811, "step": 391 }, { "epoch": 1.1895685158843055, "grad_norm": 2.1936960220336914, "learning_rate": 1.509163216996572e-05, "loss": 0.7785, "step": 392 }, { "epoch": 1.1926031294452346, "grad_norm": 2.166518449783325, "learning_rate": 1.5061151332386565e-05, "loss": 0.7775, "step": 393 }, { "epoch": 1.195637743006164, "grad_norm": 2.0909955501556396, "learning_rate": 1.5030607148298697e-05, "loss": 0.7783, "step": 394 }, { "epoch": 1.1986723565670934, "grad_norm": 2.287322521209717, "learning_rate": 1.5000000000000002e-05, "loss": 0.7678, "step": 395 }, { "epoch": 1.2017069701280227, "grad_norm": 2.265106678009033, "learning_rate": 1.4969330270576428e-05, "loss": 0.7772, "step": 396 }, { "epoch": 1.2047415836889521, "grad_norm": 2.2475264072418213, "learning_rate": 1.4938598343897215e-05, "loss": 0.7509, "step": 397 }, { "epoch": 1.2077761972498815, "grad_norm": 2.3240981101989746, "learning_rate": 1.4907804604610064e-05, "loss": 0.7849, "step": 398 }, { "epoch": 1.2108108108108109, "grad_norm": 2.148869276046753, "learning_rate": 1.4876949438136348e-05, "loss": 0.7781, "step": 399 }, { "epoch": 1.2138454243717403, "grad_norm": 2.318875789642334, "learning_rate": 1.484603323066627e-05, "loss": 0.7987, "step": 400 }, { "epoch": 1.2138454243717403, "eval_loss": NaN, "eval_runtime": 204.8025, "eval_samples_per_second": 20.302, "eval_steps_per_second": 5.078, "step": 400 }, { "epoch": 1.2168800379326696, "grad_norm": 2.1469035148620605, "learning_rate": 1.4815056369154039e-05, "loss": 0.7924, "step": 401 }, { "epoch": 1.2199146514935988, "grad_norm": 2.3983654975891113, "learning_rate": 1.4784019241313025e-05, "loss": 0.7431, "step": 402 }, { "epoch": 1.2229492650545282, "grad_norm": 2.1171765327453613, "learning_rate": 1.47529222356109e-05, "loss": 0.7583, "step": 403 }, { "epoch": 1.2259838786154575, "grad_norm": 2.3186557292938232, "learning_rate": 1.4721765741264786e-05, "loss": 0.7545, "step": 404 }, { "epoch": 1.229018492176387, "grad_norm": 2.308945417404175, "learning_rate": 1.4690550148236371e-05, "loss": 0.7752, "step": 405 }, { "epoch": 1.2320531057373163, "grad_norm": 2.141418933868408, "learning_rate": 1.4659275847227044e-05, "loss": 0.7501, "step": 406 }, { "epoch": 1.2350877192982457, "grad_norm": 2.1447696685791016, "learning_rate": 1.4627943229672992e-05, "loss": 0.7446, "step": 407 }, { "epoch": 1.238122332859175, "grad_norm": 2.062683582305908, "learning_rate": 1.4596552687740304e-05, "loss": 0.7729, "step": 408 }, { "epoch": 1.2411569464201042, "grad_norm": 2.283247232437134, "learning_rate": 1.4565104614320065e-05, "loss": 0.7752, "step": 409 }, { "epoch": 1.2441915599810336, "grad_norm": 2.299151659011841, "learning_rate": 1.453359940302344e-05, "loss": 0.7794, "step": 410 }, { "epoch": 1.247226173541963, "grad_norm": 2.2340760231018066, "learning_rate": 1.4502037448176734e-05, "loss": 0.7811, "step": 411 }, { "epoch": 1.2502607871028923, "grad_norm": 2.305233955383301, "learning_rate": 1.4470419144816483e-05, "loss": 0.7663, "step": 412 }, { "epoch": 1.2532954006638217, "grad_norm": 2.1460888385772705, "learning_rate": 1.4438744888684481e-05, "loss": 0.7584, "step": 413 }, { "epoch": 1.256330014224751, "grad_norm": 2.3122851848602295, "learning_rate": 1.4407015076222845e-05, "loss": 0.7817, "step": 414 }, { "epoch": 1.2593646277856805, "grad_norm": 2.0583643913269043, "learning_rate": 1.4375230104569044e-05, "loss": 0.7695, "step": 415 }, { "epoch": 1.2623992413466096, "grad_norm": 2.262274980545044, "learning_rate": 1.4343390371550936e-05, "loss": 0.7739, "step": 416 }, { "epoch": 1.2654338549075392, "grad_norm": 2.151893377304077, "learning_rate": 1.4311496275681785e-05, "loss": 0.7789, "step": 417 }, { "epoch": 1.2684684684684684, "grad_norm": 2.28367280960083, "learning_rate": 1.4279548216155265e-05, "loss": 0.775, "step": 418 }, { "epoch": 1.2715030820293978, "grad_norm": 2.3059751987457275, "learning_rate": 1.424754659284048e-05, "loss": 0.7613, "step": 419 }, { "epoch": 1.2745376955903271, "grad_norm": 2.436896800994873, "learning_rate": 1.4215491806276944e-05, "loss": 0.7835, "step": 420 }, { "epoch": 1.2775723091512565, "grad_norm": 2.102220058441162, "learning_rate": 1.418338425766958e-05, "loss": 0.7932, "step": 421 }, { "epoch": 1.280606922712186, "grad_norm": 2.1711723804473877, "learning_rate": 1.4151224348883692e-05, "loss": 0.7668, "step": 422 }, { "epoch": 1.2836415362731153, "grad_norm": 2.213289737701416, "learning_rate": 1.4119012482439929e-05, "loss": 0.7745, "step": 423 }, { "epoch": 1.2866761498340447, "grad_norm": 2.0960137844085693, "learning_rate": 1.408674906150926e-05, "loss": 0.7742, "step": 424 }, { "epoch": 1.2897107633949738, "grad_norm": 2.081200122833252, "learning_rate": 1.4054434489907916e-05, "loss": 0.7652, "step": 425 }, { "epoch": 1.2927453769559032, "grad_norm": 2.284423351287842, "learning_rate": 1.4022069172092354e-05, "loss": 0.7762, "step": 426 }, { "epoch": 1.2957799905168326, "grad_norm": 2.292639970779419, "learning_rate": 1.3989653513154165e-05, "loss": 0.7644, "step": 427 }, { "epoch": 1.298814604077762, "grad_norm": 2.282759428024292, "learning_rate": 1.3957187918815032e-05, "loss": 0.7658, "step": 428 }, { "epoch": 1.3018492176386913, "grad_norm": 2.2170190811157227, "learning_rate": 1.3924672795421638e-05, "loss": 0.7661, "step": 429 }, { "epoch": 1.3048838311996207, "grad_norm": 2.202991247177124, "learning_rate": 1.3892108549940583e-05, "loss": 0.7881, "step": 430 }, { "epoch": 1.30791844476055, "grad_norm": 2.148986339569092, "learning_rate": 1.3859495589953289e-05, "loss": 0.7865, "step": 431 }, { "epoch": 1.3109530583214792, "grad_norm": 2.1265058517456055, "learning_rate": 1.3826834323650899e-05, "loss": 0.789, "step": 432 }, { "epoch": 1.3139876718824088, "grad_norm": 2.1798737049102783, "learning_rate": 1.3794125159829173e-05, "loss": 0.7707, "step": 433 }, { "epoch": 1.317022285443338, "grad_norm": 2.2072978019714355, "learning_rate": 1.376136850788336e-05, "loss": 0.763, "step": 434 }, { "epoch": 1.3200568990042674, "grad_norm": 2.12349534034729, "learning_rate": 1.3728564777803089e-05, "loss": 0.7505, "step": 435 }, { "epoch": 1.3230915125651967, "grad_norm": 2.104276180267334, "learning_rate": 1.3695714380167221e-05, "loss": 0.7891, "step": 436 }, { "epoch": 1.3261261261261261, "grad_norm": 2.038515090942383, "learning_rate": 1.3662817726138729e-05, "loss": 0.7668, "step": 437 }, { "epoch": 1.3291607396870555, "grad_norm": 2.0557668209075928, "learning_rate": 1.3629875227459532e-05, "loss": 0.7685, "step": 438 }, { "epoch": 1.3321953532479849, "grad_norm": 2.221299886703491, "learning_rate": 1.359688729644536e-05, "loss": 0.7765, "step": 439 }, { "epoch": 1.3352299668089143, "grad_norm": 2.383873462677002, "learning_rate": 1.356385434598057e-05, "loss": 0.7863, "step": 440 }, { "epoch": 1.3382645803698434, "grad_norm": 2.144969940185547, "learning_rate": 1.3530776789513009e-05, "loss": 0.7854, "step": 441 }, { "epoch": 1.341299193930773, "grad_norm": 2.3431177139282227, "learning_rate": 1.3497655041048812e-05, "loss": 0.7491, "step": 442 }, { "epoch": 1.3443338074917022, "grad_norm": 2.1558756828308105, "learning_rate": 1.3464489515147239e-05, "loss": 0.7935, "step": 443 }, { "epoch": 1.3473684210526315, "grad_norm": 2.4032328128814697, "learning_rate": 1.3431280626915466e-05, "loss": 0.765, "step": 444 }, { "epoch": 1.350403034613561, "grad_norm": 2.373549461364746, "learning_rate": 1.3398028792003413e-05, "loss": 0.766, "step": 445 }, { "epoch": 1.3534376481744903, "grad_norm": 2.38440203666687, "learning_rate": 1.3364734426598527e-05, "loss": 0.7849, "step": 446 }, { "epoch": 1.3564722617354197, "grad_norm": 2.319101333618164, "learning_rate": 1.3331397947420578e-05, "loss": 0.7738, "step": 447 }, { "epoch": 1.359506875296349, "grad_norm": 2.1911866664886475, "learning_rate": 1.3298019771716435e-05, "loss": 0.7779, "step": 448 }, { "epoch": 1.3625414888572784, "grad_norm": 2.273451328277588, "learning_rate": 1.3264600317254854e-05, "loss": 0.76, "step": 449 }, { "epoch": 1.3655761024182076, "grad_norm": 2.2576355934143066, "learning_rate": 1.3231140002321252e-05, "loss": 0.7687, "step": 450 }, { "epoch": 1.3655761024182076, "eval_loss": NaN, "eval_runtime": 203.9814, "eval_samples_per_second": 20.384, "eval_steps_per_second": 5.099, "step": 450 }, { "epoch": 1.368610715979137, "grad_norm": 2.2846243381500244, "learning_rate": 1.3197639245712454e-05, "loss": 0.77, "step": 451 }, { "epoch": 1.3716453295400663, "grad_norm": 2.1583635807037354, "learning_rate": 1.3164098466731467e-05, "loss": 0.7681, "step": 452 }, { "epoch": 1.3746799431009957, "grad_norm": 2.126995325088501, "learning_rate": 1.3130518085182224e-05, "loss": 0.7755, "step": 453 }, { "epoch": 1.377714556661925, "grad_norm": 2.2357747554779053, "learning_rate": 1.3096898521364338e-05, "loss": 0.7509, "step": 454 }, { "epoch": 1.3807491702228545, "grad_norm": 2.0657386779785156, "learning_rate": 1.3063240196067837e-05, "loss": 0.8043, "step": 455 }, { "epoch": 1.3837837837837839, "grad_norm": 2.1653802394866943, "learning_rate": 1.3029543530567884e-05, "loss": 0.7676, "step": 456 }, { "epoch": 1.386818397344713, "grad_norm": 2.161508798599243, "learning_rate": 1.2995808946619533e-05, "loss": 0.7735, "step": 457 }, { "epoch": 1.3898530109056426, "grad_norm": 2.185350179672241, "learning_rate": 1.2962036866452423e-05, "loss": 0.7891, "step": 458 }, { "epoch": 1.3928876244665718, "grad_norm": 2.198807954788208, "learning_rate": 1.2928227712765504e-05, "loss": 0.7657, "step": 459 }, { "epoch": 1.3959222380275011, "grad_norm": 2.2574117183685303, "learning_rate": 1.2894381908721757e-05, "loss": 0.7893, "step": 460 }, { "epoch": 1.3989568515884305, "grad_norm": 2.2577221393585205, "learning_rate": 1.2860499877942876e-05, "loss": 0.7787, "step": 461 }, { "epoch": 1.40199146514936, "grad_norm": 2.265421152114868, "learning_rate": 1.282658204450398e-05, "loss": 0.7714, "step": 462 }, { "epoch": 1.4050260787102893, "grad_norm": 2.0731489658355713, "learning_rate": 1.2792628832928302e-05, "loss": 0.782, "step": 463 }, { "epoch": 1.4080606922712187, "grad_norm": 2.130533218383789, "learning_rate": 1.275864066818188e-05, "loss": 0.7873, "step": 464 }, { "epoch": 1.411095305832148, "grad_norm": 2.0858917236328125, "learning_rate": 1.2724617975668229e-05, "loss": 0.8005, "step": 465 }, { "epoch": 1.4141299193930772, "grad_norm": 2.0890703201293945, "learning_rate": 1.2690561181223024e-05, "loss": 0.775, "step": 466 }, { "epoch": 1.4171645329540066, "grad_norm": 2.3934412002563477, "learning_rate": 1.2656470711108763e-05, "loss": 0.7931, "step": 467 }, { "epoch": 1.420199146514936, "grad_norm": 2.2045512199401855, "learning_rate": 1.2622346992009447e-05, "loss": 0.7747, "step": 468 }, { "epoch": 1.4232337600758653, "grad_norm": 2.371346950531006, "learning_rate": 1.2588190451025209e-05, "loss": 0.7581, "step": 469 }, { "epoch": 1.4262683736367947, "grad_norm": 2.1523914337158203, "learning_rate": 1.2554001515667009e-05, "loss": 0.7741, "step": 470 }, { "epoch": 1.429302987197724, "grad_norm": 2.2948853969573975, "learning_rate": 1.2519780613851254e-05, "loss": 0.7925, "step": 471 }, { "epoch": 1.4323376007586535, "grad_norm": 2.2751922607421875, "learning_rate": 1.2485528173894447e-05, "loss": 0.7784, "step": 472 }, { "epoch": 1.4353722143195826, "grad_norm": 2.1596484184265137, "learning_rate": 1.2451244624507831e-05, "loss": 0.7895, "step": 473 }, { "epoch": 1.4384068278805122, "grad_norm": 2.2292542457580566, "learning_rate": 1.2416930394792026e-05, "loss": 0.7698, "step": 474 }, { "epoch": 1.4414414414414414, "grad_norm": 2.1559245586395264, "learning_rate": 1.238258591423165e-05, "loss": 0.776, "step": 475 }, { "epoch": 1.4444760550023708, "grad_norm": 2.072768211364746, "learning_rate": 1.234821161268995e-05, "loss": 0.7665, "step": 476 }, { "epoch": 1.4475106685633001, "grad_norm": 2.3749032020568848, "learning_rate": 1.2313807920403419e-05, "loss": 0.7765, "step": 477 }, { "epoch": 1.4505452821242295, "grad_norm": 2.1614534854888916, "learning_rate": 1.22793752679764e-05, "loss": 0.7908, "step": 478 }, { "epoch": 1.4535798956851589, "grad_norm": 2.2183918952941895, "learning_rate": 1.2244914086375726e-05, "loss": 0.7662, "step": 479 }, { "epoch": 1.4566145092460883, "grad_norm": 2.230243444442749, "learning_rate": 1.22104248069253e-05, "loss": 0.7758, "step": 480 }, { "epoch": 1.4596491228070176, "grad_norm": 2.503591775894165, "learning_rate": 1.2175907861300698e-05, "loss": 0.7739, "step": 481 }, { "epoch": 1.4626837363679468, "grad_norm": 2.4481165409088135, "learning_rate": 1.2141363681523777e-05, "loss": 0.788, "step": 482 }, { "epoch": 1.4657183499288762, "grad_norm": 2.30068302154541, "learning_rate": 1.2106792699957264e-05, "loss": 0.7905, "step": 483 }, { "epoch": 1.4687529634898056, "grad_norm": 2.2382349967956543, "learning_rate": 1.2072195349299344e-05, "loss": 0.7617, "step": 484 }, { "epoch": 1.471787577050735, "grad_norm": 2.2054882049560547, "learning_rate": 1.2037572062578238e-05, "loss": 0.7802, "step": 485 }, { "epoch": 1.4748221906116643, "grad_norm": 2.282318115234375, "learning_rate": 1.2002923273146793e-05, "loss": 0.7711, "step": 486 }, { "epoch": 1.4778568041725937, "grad_norm": 2.1077611446380615, "learning_rate": 1.1968249414677055e-05, "loss": 0.761, "step": 487 }, { "epoch": 1.480891417733523, "grad_norm": 2.1558871269226074, "learning_rate": 1.1933550921154836e-05, "loss": 0.7872, "step": 488 }, { "epoch": 1.4839260312944522, "grad_norm": 2.332897901535034, "learning_rate": 1.1898828226874284e-05, "loss": 0.7791, "step": 489 }, { "epoch": 1.4869606448553818, "grad_norm": 2.0427420139312744, "learning_rate": 1.1864081766432457e-05, "loss": 0.782, "step": 490 }, { "epoch": 1.489995258416311, "grad_norm": 2.2689926624298096, "learning_rate": 1.1829311974723868e-05, "loss": 0.7659, "step": 491 }, { "epoch": 1.4930298719772404, "grad_norm": 2.2432594299316406, "learning_rate": 1.1794519286935056e-05, "loss": 0.7543, "step": 492 }, { "epoch": 1.4960644855381697, "grad_norm": 2.0946452617645264, "learning_rate": 1.1759704138539121e-05, "loss": 0.7712, "step": 493 }, { "epoch": 1.499099099099099, "grad_norm": 2.1154541969299316, "learning_rate": 1.1724866965290302e-05, "loss": 0.7732, "step": 494 }, { "epoch": 1.5021337126600285, "grad_norm": 2.195223569869995, "learning_rate": 1.1690008203218493e-05, "loss": 0.7596, "step": 495 }, { "epoch": 1.5051683262209576, "grad_norm": 2.105151653289795, "learning_rate": 1.1655128288623803e-05, "loss": 0.775, "step": 496 }, { "epoch": 1.5082029397818872, "grad_norm": 2.196159601211548, "learning_rate": 1.1620227658071088e-05, "loss": 0.7893, "step": 497 }, { "epoch": 1.5112375533428164, "grad_norm": 2.3409769535064697, "learning_rate": 1.158530674838449e-05, "loss": 0.7494, "step": 498 }, { "epoch": 1.514272166903746, "grad_norm": 2.177128553390503, "learning_rate": 1.155036599664198e-05, "loss": 0.7595, "step": 499 }, { "epoch": 1.5173067804646752, "grad_norm": 2.3704683780670166, "learning_rate": 1.1515405840169861e-05, "loss": 0.7607, "step": 500 }, { "epoch": 1.5173067804646752, "eval_loss": NaN, "eval_runtime": 204.5697, "eval_samples_per_second": 20.326, "eval_steps_per_second": 5.084, "step": 500 }, { "epoch": 1.5203413940256045, "grad_norm": 2.175325393676758, "learning_rate": 1.1480426716537316e-05, "loss": 0.7626, "step": 501 }, { "epoch": 1.523376007586534, "grad_norm": 2.093395471572876, "learning_rate": 1.1445429063550925e-05, "loss": 0.7787, "step": 502 }, { "epoch": 1.5264106211474633, "grad_norm": 2.2766175270080566, "learning_rate": 1.1410413319249193e-05, "loss": 0.7592, "step": 503 }, { "epoch": 1.5294452347083927, "grad_norm": 2.00833797454834, "learning_rate": 1.1375379921897052e-05, "loss": 0.7685, "step": 504 }, { "epoch": 1.5324798482693218, "grad_norm": 2.2018191814422607, "learning_rate": 1.1340329309980379e-05, "loss": 0.753, "step": 505 }, { "epoch": 1.5355144618302514, "grad_norm": 2.228724479675293, "learning_rate": 1.130526192220052e-05, "loss": 0.7687, "step": 506 }, { "epoch": 1.5385490753911806, "grad_norm": 2.241725444793701, "learning_rate": 1.1270178197468788e-05, "loss": 0.7674, "step": 507 }, { "epoch": 1.54158368895211, "grad_norm": 2.1697564125061035, "learning_rate": 1.1235078574900984e-05, "loss": 0.7726, "step": 508 }, { "epoch": 1.5446183025130393, "grad_norm": 2.197449207305908, "learning_rate": 1.119996349381187e-05, "loss": 0.7672, "step": 509 }, { "epoch": 1.5476529160739687, "grad_norm": 2.0595834255218506, "learning_rate": 1.1164833393709707e-05, "loss": 0.7706, "step": 510 }, { "epoch": 1.550687529634898, "grad_norm": 2.3002591133117676, "learning_rate": 1.112968871429073e-05, "loss": 0.7875, "step": 511 }, { "epoch": 1.5537221431958272, "grad_norm": 2.175219774246216, "learning_rate": 1.1094529895433653e-05, "loss": 0.7809, "step": 512 }, { "epoch": 1.5567567567567568, "grad_norm": 2.1368846893310547, "learning_rate": 1.1059357377194161e-05, "loss": 0.7878, "step": 513 }, { "epoch": 1.559791370317686, "grad_norm": 2.210344076156616, "learning_rate": 1.102417159979941e-05, "loss": 0.7543, "step": 514 }, { "epoch": 1.5628259838786156, "grad_norm": 2.1888577938079834, "learning_rate": 1.09889730036425e-05, "loss": 0.7731, "step": 515 }, { "epoch": 1.5658605974395448, "grad_norm": 2.3701112270355225, "learning_rate": 1.0953762029276982e-05, "loss": 0.7689, "step": 516 }, { "epoch": 1.5688952110004741, "grad_norm": 2.069556713104248, "learning_rate": 1.0918539117411334e-05, "loss": 0.767, "step": 517 }, { "epoch": 1.5719298245614035, "grad_norm": 2.209773302078247, "learning_rate": 1.0883304708903441e-05, "loss": 0.7696, "step": 518 }, { "epoch": 1.5749644381223329, "grad_norm": 2.1156795024871826, "learning_rate": 1.0848059244755093e-05, "loss": 0.7827, "step": 519 }, { "epoch": 1.5779990516832623, "grad_norm": 2.359513521194458, "learning_rate": 1.0812803166106445e-05, "loss": 0.7612, "step": 520 }, { "epoch": 1.5810336652441914, "grad_norm": 2.2499759197235107, "learning_rate": 1.0777536914230509e-05, "loss": 0.7671, "step": 521 }, { "epoch": 1.584068278805121, "grad_norm": 2.219525098800659, "learning_rate": 1.0742260930527625e-05, "loss": 0.776, "step": 522 }, { "epoch": 1.5871028923660502, "grad_norm": 2.339210033416748, "learning_rate": 1.0706975656519946e-05, "loss": 0.7669, "step": 523 }, { "epoch": 1.5901375059269798, "grad_norm": 2.3282480239868164, "learning_rate": 1.06716815338459e-05, "loss": 0.7843, "step": 524 }, { "epoch": 1.593172119487909, "grad_norm": 2.09635329246521, "learning_rate": 1.0636379004254665e-05, "loss": 0.7598, "step": 525 }, { "epoch": 1.5962067330488383, "grad_norm": 2.3128199577331543, "learning_rate": 1.0601068509600642e-05, "loss": 0.7673, "step": 526 }, { "epoch": 1.5992413466097677, "grad_norm": 2.3669700622558594, "learning_rate": 1.0565750491837925e-05, "loss": 0.7697, "step": 527 }, { "epoch": 1.602275960170697, "grad_norm": 2.2540953159332275, "learning_rate": 1.0530425393014773e-05, "loss": 0.7641, "step": 528 }, { "epoch": 1.6053105737316264, "grad_norm": 2.3089439868927, "learning_rate": 1.049509365526807e-05, "loss": 0.768, "step": 529 }, { "epoch": 1.6083451872925556, "grad_norm": 2.0084259510040283, "learning_rate": 1.0459755720817797e-05, "loss": 0.7504, "step": 530 }, { "epoch": 1.6113798008534852, "grad_norm": 2.261798143386841, "learning_rate": 1.0424412031961485e-05, "loss": 0.7629, "step": 531 }, { "epoch": 1.6144144144144144, "grad_norm": 2.2108161449432373, "learning_rate": 1.0389063031068698e-05, "loss": 0.738, "step": 532 }, { "epoch": 1.6174490279753437, "grad_norm": 2.2087652683258057, "learning_rate": 1.0353709160575488e-05, "loss": 0.7758, "step": 533 }, { "epoch": 1.6204836415362731, "grad_norm": 2.218838691711426, "learning_rate": 1.0318350862978848e-05, "loss": 0.7641, "step": 534 }, { "epoch": 1.6235182550972025, "grad_norm": 2.156991720199585, "learning_rate": 1.0282988580831183e-05, "loss": 0.7577, "step": 535 }, { "epoch": 1.6265528686581319, "grad_norm": 2.1950886249542236, "learning_rate": 1.0247622756734775e-05, "loss": 0.7888, "step": 536 }, { "epoch": 1.629587482219061, "grad_norm": 2.114649534225464, "learning_rate": 1.0212253833336237e-05, "loss": 0.7766, "step": 537 }, { "epoch": 1.6326220957799906, "grad_norm": 2.217928409576416, "learning_rate": 1.0176882253320968e-05, "loss": 0.7529, "step": 538 }, { "epoch": 1.6356567093409198, "grad_norm": 2.2367630004882812, "learning_rate": 1.0141508459407622e-05, "loss": 0.7699, "step": 539 }, { "epoch": 1.6386913229018494, "grad_norm": 2.1163787841796875, "learning_rate": 1.0106132894342564e-05, "loss": 0.7637, "step": 540 }, { "epoch": 1.6417259364627785, "grad_norm": 2.083292245864868, "learning_rate": 1.0070756000894321e-05, "loss": 0.783, "step": 541 }, { "epoch": 1.644760550023708, "grad_norm": 2.1091578006744385, "learning_rate": 1.0035378221848053e-05, "loss": 0.7848, "step": 542 }, { "epoch": 1.6477951635846373, "grad_norm": 2.0994367599487305, "learning_rate": 1e-05, "loss": 0.7698, "step": 543 }, { "epoch": 1.6508297771455667, "grad_norm": 2.2539420127868652, "learning_rate": 9.964621778151947e-06, "loss": 0.769, "step": 544 }, { "epoch": 1.653864390706496, "grad_norm": 2.154353380203247, "learning_rate": 9.929243999105682e-06, "loss": 0.7496, "step": 545 }, { "epoch": 1.6568990042674252, "grad_norm": 2.265385150909424, "learning_rate": 9.89386710565744e-06, "loss": 0.768, "step": 546 }, { "epoch": 1.6599336178283548, "grad_norm": 2.073289632797241, "learning_rate": 9.858491540592383e-06, "loss": 0.7807, "step": 547 }, { "epoch": 1.662968231389284, "grad_norm": 2.135737657546997, "learning_rate": 9.823117746679034e-06, "loss": 0.766, "step": 548 }, { "epoch": 1.6660028449502133, "grad_norm": 2.1423635482788086, "learning_rate": 9.787746166663765e-06, "loss": 0.7588, "step": 549 }, { "epoch": 1.6690374585111427, "grad_norm": 2.0772769451141357, "learning_rate": 9.752377243265229e-06, "loss": 0.7647, "step": 550 }, { "epoch": 1.6690374585111427, "eval_loss": NaN, "eval_runtime": 204.4783, "eval_samples_per_second": 20.335, "eval_steps_per_second": 5.086, "step": 550 }, { "epoch": 1.672072072072072, "grad_norm": 2.118854284286499, "learning_rate": 9.71701141916882e-06, "loss": 0.7728, "step": 551 }, { "epoch": 1.6751066856330015, "grad_norm": 2.1113169193267822, "learning_rate": 9.681649137021158e-06, "loss": 0.7796, "step": 552 }, { "epoch": 1.6781412991939306, "grad_norm": 2.064993143081665, "learning_rate": 9.646290839424515e-06, "loss": 0.762, "step": 553 }, { "epoch": 1.6811759127548602, "grad_norm": 2.151028633117676, "learning_rate": 9.610936968931302e-06, "loss": 0.7709, "step": 554 }, { "epoch": 1.6842105263157894, "grad_norm": 2.201429843902588, "learning_rate": 9.57558796803852e-06, "loss": 0.7619, "step": 555 }, { "epoch": 1.687245139876719, "grad_norm": 2.065720796585083, "learning_rate": 9.540244279182206e-06, "loss": 0.798, "step": 556 }, { "epoch": 1.6902797534376481, "grad_norm": 2.1895217895507812, "learning_rate": 9.504906344731933e-06, "loss": 0.7802, "step": 557 }, { "epoch": 1.6933143669985775, "grad_norm": 2.2065324783325195, "learning_rate": 9.46957460698523e-06, "loss": 0.7767, "step": 558 }, { "epoch": 1.696348980559507, "grad_norm": 2.17842698097229, "learning_rate": 9.434249508162076e-06, "loss": 0.7604, "step": 559 }, { "epoch": 1.6993835941204363, "grad_norm": 2.164586305618286, "learning_rate": 9.398931490399363e-06, "loss": 0.756, "step": 560 }, { "epoch": 1.7024182076813656, "grad_norm": 2.235797643661499, "learning_rate": 9.363620995745337e-06, "loss": 0.777, "step": 561 }, { "epoch": 1.7054528212422948, "grad_norm": 2.084228038787842, "learning_rate": 9.328318466154102e-06, "loss": 0.765, "step": 562 }, { "epoch": 1.7084874348032244, "grad_norm": 2.1382224559783936, "learning_rate": 9.293024343480056e-06, "loss": 0.7712, "step": 563 }, { "epoch": 1.7115220483641536, "grad_norm": 2.304201126098633, "learning_rate": 9.257739069472375e-06, "loss": 0.7748, "step": 564 }, { "epoch": 1.714556661925083, "grad_norm": 2.1342921257019043, "learning_rate": 9.222463085769495e-06, "loss": 0.762, "step": 565 }, { "epoch": 1.7175912754860123, "grad_norm": 2.189845323562622, "learning_rate": 9.187196833893559e-06, "loss": 0.7605, "step": 566 }, { "epoch": 1.7206258890469417, "grad_norm": 2.3078930377960205, "learning_rate": 9.151940755244912e-06, "loss": 0.7677, "step": 567 }, { "epoch": 1.723660502607871, "grad_norm": 2.303213357925415, "learning_rate": 9.11669529109656e-06, "loss": 0.7735, "step": 568 }, { "epoch": 1.7266951161688002, "grad_norm": 2.3082435131073, "learning_rate": 9.081460882588668e-06, "loss": 0.7683, "step": 569 }, { "epoch": 1.7297297297297298, "grad_norm": 2.2995338439941406, "learning_rate": 9.046237970723022e-06, "loss": 0.752, "step": 570 }, { "epoch": 1.732764343290659, "grad_norm": 2.0647215843200684, "learning_rate": 9.011026996357504e-06, "loss": 0.7731, "step": 571 }, { "epoch": 1.7357989568515886, "grad_norm": 2.160496473312378, "learning_rate": 8.975828400200592e-06, "loss": 0.7543, "step": 572 }, { "epoch": 1.7388335704125177, "grad_norm": 2.3624351024627686, "learning_rate": 8.94064262280584e-06, "loss": 0.774, "step": 573 }, { "epoch": 1.7418681839734471, "grad_norm": 2.165922164916992, "learning_rate": 8.90547010456635e-06, "loss": 0.7815, "step": 574 }, { "epoch": 1.7449027975343765, "grad_norm": 2.138921022415161, "learning_rate": 8.870311285709274e-06, "loss": 0.7593, "step": 575 }, { "epoch": 1.7479374110953059, "grad_norm": 2.122746706008911, "learning_rate": 8.835166606290295e-06, "loss": 0.783, "step": 576 }, { "epoch": 1.7509720246562352, "grad_norm": 2.119318723678589, "learning_rate": 8.80003650618813e-06, "loss": 0.7778, "step": 577 }, { "epoch": 1.7540066382171644, "grad_norm": 2.092092990875244, "learning_rate": 8.76492142509902e-06, "loss": 0.765, "step": 578 }, { "epoch": 1.757041251778094, "grad_norm": 2.156517267227173, "learning_rate": 8.729821802531213e-06, "loss": 0.7583, "step": 579 }, { "epoch": 1.7600758653390232, "grad_norm": 2.2096939086914062, "learning_rate": 8.694738077799487e-06, "loss": 0.7698, "step": 580 }, { "epoch": 1.7631104788999525, "grad_norm": 1.981307029724121, "learning_rate": 8.659670690019626e-06, "loss": 0.7771, "step": 581 }, { "epoch": 1.766145092460882, "grad_norm": 2.1369874477386475, "learning_rate": 8.624620078102952e-06, "loss": 0.777, "step": 582 }, { "epoch": 1.7691797060218113, "grad_norm": 2.1329455375671387, "learning_rate": 8.58958668075081e-06, "loss": 0.7748, "step": 583 }, { "epoch": 1.7722143195827407, "grad_norm": 2.1630642414093018, "learning_rate": 8.554570936449073e-06, "loss": 0.7675, "step": 584 }, { "epoch": 1.7752489331436698, "grad_norm": 2.1383938789367676, "learning_rate": 8.519573283462688e-06, "loss": 0.7702, "step": 585 }, { "epoch": 1.7782835467045994, "grad_norm": 2.001569986343384, "learning_rate": 8.484594159830142e-06, "loss": 0.7553, "step": 586 }, { "epoch": 1.7813181602655286, "grad_norm": 2.1266000270843506, "learning_rate": 8.449634003358022e-06, "loss": 0.7757, "step": 587 }, { "epoch": 1.7843527738264582, "grad_norm": 2.172466278076172, "learning_rate": 8.414693251615513e-06, "loss": 0.7625, "step": 588 }, { "epoch": 1.7873873873873873, "grad_norm": 2.2140755653381348, "learning_rate": 8.379772341928916e-06, "loss": 0.7653, "step": 589 }, { "epoch": 1.7904220009483167, "grad_norm": 2.1517114639282227, "learning_rate": 8.344871711376202e-06, "loss": 0.7744, "step": 590 }, { "epoch": 1.793456614509246, "grad_norm": 2.1749329566955566, "learning_rate": 8.309991796781512e-06, "loss": 0.7663, "step": 591 }, { "epoch": 1.7964912280701755, "grad_norm": 2.2055773735046387, "learning_rate": 8.2751330347097e-06, "loss": 0.7768, "step": 592 }, { "epoch": 1.7995258416311048, "grad_norm": 2.117532253265381, "learning_rate": 8.24029586146088e-06, "loss": 0.7774, "step": 593 }, { "epoch": 1.802560455192034, "grad_norm": 2.2061290740966797, "learning_rate": 8.205480713064947e-06, "loss": 0.7819, "step": 594 }, { "epoch": 1.8055950687529636, "grad_norm": 2.2168619632720947, "learning_rate": 8.170688025276134e-06, "loss": 0.7646, "step": 595 }, { "epoch": 1.8086296823138928, "grad_norm": 2.238391637802124, "learning_rate": 8.135918233567545e-06, "loss": 0.7782, "step": 596 }, { "epoch": 1.8116642958748221, "grad_norm": 2.105041265487671, "learning_rate": 8.101171773125716e-06, "loss": 0.7525, "step": 597 }, { "epoch": 1.8146989094357515, "grad_norm": 2.0234577655792236, "learning_rate": 8.066449078845168e-06, "loss": 0.7658, "step": 598 }, { "epoch": 1.817733522996681, "grad_norm": 2.188751220703125, "learning_rate": 8.031750585322948e-06, "loss": 0.7616, "step": 599 }, { "epoch": 1.8207681365576103, "grad_norm": 2.1460800170898438, "learning_rate": 7.99707672685321e-06, "loss": 0.7617, "step": 600 }, { "epoch": 1.8207681365576103, "eval_loss": NaN, "eval_runtime": 204.6188, "eval_samples_per_second": 20.321, "eval_steps_per_second": 5.083, "step": 600 }, { "epoch": 1.8238027501185394, "grad_norm": 1.964609146118164, "learning_rate": 7.962427937421763e-06, "loss": 0.7876, "step": 601 }, { "epoch": 1.826837363679469, "grad_norm": 2.0918426513671875, "learning_rate": 7.92780465070066e-06, "loss": 0.7744, "step": 602 }, { "epoch": 1.8298719772403982, "grad_norm": 2.059992790222168, "learning_rate": 7.89320730004274e-06, "loss": 0.7665, "step": 603 }, { "epoch": 1.8329065908013278, "grad_norm": 2.0450618267059326, "learning_rate": 7.858636318476226e-06, "loss": 0.7636, "step": 604 }, { "epoch": 1.835941204362257, "grad_norm": 2.1291024684906006, "learning_rate": 7.824092138699307e-06, "loss": 0.7652, "step": 605 }, { "epoch": 1.8389758179231863, "grad_norm": 2.0631322860717773, "learning_rate": 7.789575193074703e-06, "loss": 0.7819, "step": 606 }, { "epoch": 1.8420104314841157, "grad_norm": 2.1132242679595947, "learning_rate": 7.755085913624274e-06, "loss": 0.7507, "step": 607 }, { "epoch": 1.845045045045045, "grad_norm": 2.1836376190185547, "learning_rate": 7.720624732023604e-06, "loss": 0.7798, "step": 608 }, { "epoch": 1.8480796586059745, "grad_norm": 2.087453842163086, "learning_rate": 7.686192079596586e-06, "loss": 0.7817, "step": 609 }, { "epoch": 1.8511142721669036, "grad_norm": 2.0749568939208984, "learning_rate": 7.651788387310053e-06, "loss": 0.7893, "step": 610 }, { "epoch": 1.8541488857278332, "grad_norm": 2.085495710372925, "learning_rate": 7.617414085768352e-06, "loss": 0.7765, "step": 611 }, { "epoch": 1.8571834992887624, "grad_norm": 2.1710901260375977, "learning_rate": 7.5830696052079754e-06, "loss": 0.7745, "step": 612 }, { "epoch": 1.860218112849692, "grad_norm": 2.12780499458313, "learning_rate": 7.548755375492173e-06, "loss": 0.7845, "step": 613 }, { "epoch": 1.8632527264106211, "grad_norm": 2.240459680557251, "learning_rate": 7.514471826105556e-06, "loss": 0.7731, "step": 614 }, { "epoch": 1.8662873399715505, "grad_norm": 2.130253553390503, "learning_rate": 7.480219386148751e-06, "loss": 0.7904, "step": 615 }, { "epoch": 1.8693219535324799, "grad_norm": 2.216806650161743, "learning_rate": 7.445998484332994e-06, "loss": 0.7529, "step": 616 }, { "epoch": 1.8723565670934093, "grad_norm": 2.250821352005005, "learning_rate": 7.411809548974792e-06, "loss": 0.7566, "step": 617 }, { "epoch": 1.8753911806543386, "grad_norm": 2.088843822479248, "learning_rate": 7.377653007990559e-06, "loss": 0.7568, "step": 618 }, { "epoch": 1.8784257942152678, "grad_norm": 2.108356237411499, "learning_rate": 7.343529288891239e-06, "loss": 0.7769, "step": 619 }, { "epoch": 1.8814604077761974, "grad_norm": 2.1201224327087402, "learning_rate": 7.3094388187769815e-06, "loss": 0.7651, "step": 620 }, { "epoch": 1.8844950213371265, "grad_norm": 2.066652536392212, "learning_rate": 7.275382024331773e-06, "loss": 0.7773, "step": 621 }, { "epoch": 1.887529634898056, "grad_norm": 2.1200613975524902, "learning_rate": 7.241359331818121e-06, "loss": 0.7835, "step": 622 }, { "epoch": 1.8905642484589853, "grad_norm": 2.160304546356201, "learning_rate": 7.2073711670717e-06, "loss": 0.7489, "step": 623 }, { "epoch": 1.8935988620199147, "grad_norm": 2.2403597831726074, "learning_rate": 7.173417955496025e-06, "loss": 0.7579, "step": 624 }, { "epoch": 1.896633475580844, "grad_norm": 2.083029270172119, "learning_rate": 7.13950012205713e-06, "loss": 0.7715, "step": 625 }, { "epoch": 1.8996680891417732, "grad_norm": 2.031341075897217, "learning_rate": 7.105618091278246e-06, "loss": 0.7494, "step": 626 }, { "epoch": 1.9027027027027028, "grad_norm": 2.155595541000366, "learning_rate": 7.071772287234497e-06, "loss": 0.7797, "step": 627 }, { "epoch": 1.905737316263632, "grad_norm": 2.2580630779266357, "learning_rate": 7.037963133547583e-06, "loss": 0.7801, "step": 628 }, { "epoch": 1.9087719298245616, "grad_norm": 2.2416763305664062, "learning_rate": 7.004191053380469e-06, "loss": 0.7753, "step": 629 }, { "epoch": 1.9118065433854907, "grad_norm": 2.0690417289733887, "learning_rate": 6.970456469432116e-06, "loss": 0.7693, "step": 630 }, { "epoch": 1.91484115694642, "grad_norm": 2.0331270694732666, "learning_rate": 6.936759803932167e-06, "loss": 0.7578, "step": 631 }, { "epoch": 1.9178757705073495, "grad_norm": 2.1094417572021484, "learning_rate": 6.903101478635663e-06, "loss": 0.7589, "step": 632 }, { "epoch": 1.9209103840682789, "grad_norm": 2.042919635772705, "learning_rate": 6.869481914817779e-06, "loss": 0.7659, "step": 633 }, { "epoch": 1.9239449976292082, "grad_norm": 2.0610294342041016, "learning_rate": 6.835901533268536e-06, "loss": 0.7648, "step": 634 }, { "epoch": 1.9269796111901374, "grad_norm": 2.0473015308380127, "learning_rate": 6.802360754287548e-06, "loss": 0.7763, "step": 635 }, { "epoch": 1.930014224751067, "grad_norm": 2.1137821674346924, "learning_rate": 6.768859997678751e-06, "loss": 0.7849, "step": 636 }, { "epoch": 1.9330488383119961, "grad_norm": 2.12430477142334, "learning_rate": 6.735399682745145e-06, "loss": 0.7644, "step": 637 }, { "epoch": 1.9360834518729255, "grad_norm": 2.0620853900909424, "learning_rate": 6.701980228283569e-06, "loss": 0.7733, "step": 638 }, { "epoch": 1.939118065433855, "grad_norm": 2.124861478805542, "learning_rate": 6.668602052579425e-06, "loss": 0.7538, "step": 639 }, { "epoch": 1.9421526789947843, "grad_norm": 2.003215789794922, "learning_rate": 6.635265573401475e-06, "loss": 0.7574, "step": 640 }, { "epoch": 1.9451872925557137, "grad_norm": 2.1591827869415283, "learning_rate": 6.601971207996592e-06, "loss": 0.7591, "step": 641 }, { "epoch": 1.9482219061166428, "grad_norm": 2.1029608249664307, "learning_rate": 6.5687193730845375e-06, "loss": 0.7927, "step": 642 }, { "epoch": 1.9512565196775724, "grad_norm": 2.069796085357666, "learning_rate": 6.535510484852767e-06, "loss": 0.7491, "step": 643 }, { "epoch": 1.9542911332385016, "grad_norm": 2.0876927375793457, "learning_rate": 6.50234495895119e-06, "loss": 0.7615, "step": 644 }, { "epoch": 1.9573257467994312, "grad_norm": 2.0911192893981934, "learning_rate": 6.469223210486992e-06, "loss": 0.7579, "step": 645 }, { "epoch": 1.9603603603603603, "grad_norm": 2.2714855670928955, "learning_rate": 6.4361456540194325e-06, "loss": 0.7543, "step": 646 }, { "epoch": 1.9633949739212897, "grad_norm": 2.075011968612671, "learning_rate": 6.403112703554643e-06, "loss": 0.752, "step": 647 }, { "epoch": 1.966429587482219, "grad_norm": 2.097029447555542, "learning_rate": 6.370124772540469e-06, "loss": 0.7338, "step": 648 }, { "epoch": 1.9694642010431485, "grad_norm": 2.147951602935791, "learning_rate": 6.337182273861273e-06, "loss": 0.7735, "step": 649 }, { "epoch": 1.9724988146040778, "grad_norm": 2.0643298625946045, "learning_rate": 6.3042856198327795e-06, "loss": 0.7775, "step": 650 }, { "epoch": 1.9724988146040778, "eval_loss": NaN, "eval_runtime": 204.7127, "eval_samples_per_second": 20.311, "eval_steps_per_second": 5.08, "step": 650 }, { "epoch": 1.975533428165007, "grad_norm": 2.043440580368042, "learning_rate": 6.2714352221969155e-06, "loss": 0.7593, "step": 651 }, { "epoch": 1.9785680417259366, "grad_norm": 2.250242233276367, "learning_rate": 6.238631492116643e-06, "loss": 0.7657, "step": 652 }, { "epoch": 1.9816026552868657, "grad_norm": 2.068997383117676, "learning_rate": 6.205874840170833e-06, "loss": 0.7774, "step": 653 }, { "epoch": 1.9846372688477951, "grad_norm": 2.1658010482788086, "learning_rate": 6.173165676349103e-06, "loss": 0.7792, "step": 654 }, { "epoch": 1.9876718824087245, "grad_norm": 2.1947412490844727, "learning_rate": 6.140504410046712e-06, "loss": 0.7631, "step": 655 }, { "epoch": 1.9907064959696539, "grad_norm": 2.082087278366089, "learning_rate": 6.107891450059419e-06, "loss": 0.7564, "step": 656 }, { "epoch": 1.9937411095305833, "grad_norm": 2.043905019760132, "learning_rate": 6.075327204578363e-06, "loss": 0.7834, "step": 657 }, { "epoch": 1.9967757230915124, "grad_norm": 2.1089298725128174, "learning_rate": 6.042812081184973e-06, "loss": 0.7595, "step": 658 }, { "epoch": 1.999810336652442, "grad_norm": 2.0419466495513916, "learning_rate": 6.010346486845837e-06, "loss": 0.7481, "step": 659 }, { "epoch": 2.002844950213371, "grad_norm": 2.1154048442840576, "learning_rate": 5.97793082790765e-06, "loss": 0.7479, "step": 660 }, { "epoch": 2.0058795637743008, "grad_norm": 1.9878915548324585, "learning_rate": 5.945565510092086e-06, "loss": 0.7547, "step": 661 }, { "epoch": 2.00891417733523, "grad_norm": 2.1136927604675293, "learning_rate": 5.913250938490744e-06, "loss": 0.7294, "step": 662 }, { "epoch": 2.0119487908961595, "grad_norm": 2.1599817276000977, "learning_rate": 5.880987517560075e-06, "loss": 0.744, "step": 663 }, { "epoch": 2.0149834044570887, "grad_norm": 2.200779438018799, "learning_rate": 5.84877565111631e-06, "loss": 0.7353, "step": 664 }, { "epoch": 2.018018018018018, "grad_norm": 2.05635404586792, "learning_rate": 5.81661574233042e-06, "loss": 0.7534, "step": 665 }, { "epoch": 2.0210526315789474, "grad_norm": 2.193406820297241, "learning_rate": 5.784508193723058e-06, "loss": 0.7443, "step": 666 }, { "epoch": 2.0240872451398766, "grad_norm": 2.120541572570801, "learning_rate": 5.752453407159521e-06, "loss": 0.7414, "step": 667 }, { "epoch": 2.027121858700806, "grad_norm": 2.075017213821411, "learning_rate": 5.720451783844741e-06, "loss": 0.7439, "step": 668 }, { "epoch": 2.0301564722617353, "grad_norm": 2.1643104553222656, "learning_rate": 5.688503724318217e-06, "loss": 0.7565, "step": 669 }, { "epoch": 2.033191085822665, "grad_norm": 2.148364782333374, "learning_rate": 5.656609628449064e-06, "loss": 0.7353, "step": 670 }, { "epoch": 2.036225699383594, "grad_norm": 2.205618143081665, "learning_rate": 5.6247698954309616e-06, "loss": 0.7529, "step": 671 }, { "epoch": 2.0392603129445233, "grad_norm": 2.1727452278137207, "learning_rate": 5.592984923777156e-06, "loss": 0.7439, "step": 672 }, { "epoch": 2.042294926505453, "grad_norm": 2.1897435188293457, "learning_rate": 5.561255111315525e-06, "loss": 0.7578, "step": 673 }, { "epoch": 2.045329540066382, "grad_norm": 2.155768871307373, "learning_rate": 5.529580855183518e-06, "loss": 0.7687, "step": 674 }, { "epoch": 2.0483641536273116, "grad_norm": 2.100489616394043, "learning_rate": 5.497962551823266e-06, "loss": 0.7333, "step": 675 }, { "epoch": 2.0513987671882408, "grad_norm": 2.0687026977539062, "learning_rate": 5.466400596976568e-06, "loss": 0.745, "step": 676 }, { "epoch": 2.0544333807491704, "grad_norm": 2.144482135772705, "learning_rate": 5.434895385679937e-06, "loss": 0.7369, "step": 677 }, { "epoch": 2.0574679943100995, "grad_norm": 2.1690595149993896, "learning_rate": 5.403447312259702e-06, "loss": 0.7443, "step": 678 }, { "epoch": 2.060502607871029, "grad_norm": 2.0905041694641113, "learning_rate": 5.3720567703270135e-06, "loss": 0.7333, "step": 679 }, { "epoch": 2.0635372214319583, "grad_norm": 2.205411911010742, "learning_rate": 5.340724152772956e-06, "loss": 0.7399, "step": 680 }, { "epoch": 2.0665718349928874, "grad_norm": 2.3007540702819824, "learning_rate": 5.3094498517636324e-06, "loss": 0.7452, "step": 681 }, { "epoch": 2.069606448553817, "grad_norm": 2.2882180213928223, "learning_rate": 5.278234258735215e-06, "loss": 0.7285, "step": 682 }, { "epoch": 2.072641062114746, "grad_norm": 2.2169337272644043, "learning_rate": 5.247077764389099e-06, "loss": 0.7437, "step": 683 }, { "epoch": 2.075675675675676, "grad_norm": 2.1462368965148926, "learning_rate": 5.215980758686978e-06, "loss": 0.7366, "step": 684 }, { "epoch": 2.078710289236605, "grad_norm": 2.200030565261841, "learning_rate": 5.18494363084596e-06, "loss": 0.7546, "step": 685 }, { "epoch": 2.0817449027975345, "grad_norm": 2.1369688510894775, "learning_rate": 5.153966769333734e-06, "loss": 0.7388, "step": 686 }, { "epoch": 2.0847795163584637, "grad_norm": 2.191958427429199, "learning_rate": 5.1230505618636575e-06, "loss": 0.7297, "step": 687 }, { "epoch": 2.0878141299193933, "grad_norm": 2.1999082565307617, "learning_rate": 5.092195395389937e-06, "loss": 0.7428, "step": 688 }, { "epoch": 2.0908487434803225, "grad_norm": 2.373140335083008, "learning_rate": 5.061401656102791e-06, "loss": 0.7585, "step": 689 }, { "epoch": 2.0938833570412516, "grad_norm": 2.2395219802856445, "learning_rate": 5.030669729423572e-06, "loss": 0.7382, "step": 690 }, { "epoch": 2.096917970602181, "grad_norm": 2.1638119220733643, "learning_rate": 5.000000000000003e-06, "loss": 0.7538, "step": 691 }, { "epoch": 2.0999525841631104, "grad_norm": 2.173945188522339, "learning_rate": 4.969392851701306e-06, "loss": 0.7403, "step": 692 }, { "epoch": 2.10298719772404, "grad_norm": 2.2739264965057373, "learning_rate": 4.938848667613436e-06, "loss": 0.7459, "step": 693 }, { "epoch": 2.106021811284969, "grad_norm": 2.2303757667541504, "learning_rate": 4.908367830034284e-06, "loss": 0.717, "step": 694 }, { "epoch": 2.1090564248458987, "grad_norm": 2.083385944366455, "learning_rate": 4.8779507204688595e-06, "loss": 0.747, "step": 695 }, { "epoch": 2.112091038406828, "grad_norm": 2.252095937728882, "learning_rate": 4.84759771962455e-06, "loss": 0.7414, "step": 696 }, { "epoch": 2.115125651967757, "grad_norm": 2.192476272583008, "learning_rate": 4.817309207406347e-06, "loss": 0.7078, "step": 697 }, { "epoch": 2.1181602655286866, "grad_norm": 2.170762062072754, "learning_rate": 4.787085562912076e-06, "loss": 0.7482, "step": 698 }, { "epoch": 2.121194879089616, "grad_norm": 2.1275007724761963, "learning_rate": 4.756927164427685e-06, "loss": 0.7415, "step": 699 }, { "epoch": 2.1242294926505454, "grad_norm": 2.1509687900543213, "learning_rate": 4.726834389422461e-06, "loss": 0.7326, "step": 700 }, { "epoch": 2.1242294926505454, "eval_loss": NaN, "eval_runtime": 204.0742, "eval_samples_per_second": 20.375, "eval_steps_per_second": 5.096, "step": 700 }, { "epoch": 2.1272641062114745, "grad_norm": 2.0226032733917236, "learning_rate": 4.696807614544352e-06, "loss": 0.7411, "step": 701 }, { "epoch": 2.130298719772404, "grad_norm": 2.1866512298583984, "learning_rate": 4.666847215615225e-06, "loss": 0.7394, "step": 702 }, { "epoch": 2.1333333333333333, "grad_norm": 2.2013630867004395, "learning_rate": 4.636953567626176e-06, "loss": 0.7356, "step": 703 }, { "epoch": 2.1363679468942625, "grad_norm": 2.4076430797576904, "learning_rate": 4.607127044732827e-06, "loss": 0.7501, "step": 704 }, { "epoch": 2.139402560455192, "grad_norm": 2.1296753883361816, "learning_rate": 4.57736802025065e-06, "loss": 0.7399, "step": 705 }, { "epoch": 2.142437174016121, "grad_norm": 2.22512149810791, "learning_rate": 4.5476768666502895e-06, "loss": 0.7553, "step": 706 }, { "epoch": 2.145471787577051, "grad_norm": 2.178199052810669, "learning_rate": 4.518053955552903e-06, "loss": 0.7458, "step": 707 }, { "epoch": 2.14850640113798, "grad_norm": 2.0975606441497803, "learning_rate": 4.488499657725511e-06, "loss": 0.7234, "step": 708 }, { "epoch": 2.1515410146989096, "grad_norm": 2.1103413105010986, "learning_rate": 4.459014343076356e-06, "loss": 0.7431, "step": 709 }, { "epoch": 2.1545756282598387, "grad_norm": 2.2594456672668457, "learning_rate": 4.429598380650266e-06, "loss": 0.7578, "step": 710 }, { "epoch": 2.1576102418207683, "grad_norm": 2.2270963191986084, "learning_rate": 4.400252138624047e-06, "loss": 0.7461, "step": 711 }, { "epoch": 2.1606448553816975, "grad_norm": 2.191631555557251, "learning_rate": 4.370975984301866e-06, "loss": 0.7425, "step": 712 }, { "epoch": 2.1636794689426266, "grad_norm": 2.124616861343384, "learning_rate": 4.341770284110655e-06, "loss": 0.7495, "step": 713 }, { "epoch": 2.1667140825035562, "grad_norm": 2.18497371673584, "learning_rate": 4.312635403595532e-06, "loss": 0.7239, "step": 714 }, { "epoch": 2.1697486960644854, "grad_norm": 2.149658441543579, "learning_rate": 4.283571707415214e-06, "loss": 0.7325, "step": 715 }, { "epoch": 2.172783309625415, "grad_norm": 2.133171558380127, "learning_rate": 4.25457955933746e-06, "loss": 0.7324, "step": 716 }, { "epoch": 2.175817923186344, "grad_norm": 2.15606689453125, "learning_rate": 4.2256593222345185e-06, "loss": 0.7273, "step": 717 }, { "epoch": 2.1788525367472737, "grad_norm": 2.237285614013672, "learning_rate": 4.196811358078585e-06, "loss": 0.7463, "step": 718 }, { "epoch": 2.181887150308203, "grad_norm": 2.267974853515625, "learning_rate": 4.168036027937267e-06, "loss": 0.7405, "step": 719 }, { "epoch": 2.1849217638691325, "grad_norm": 2.0772793292999268, "learning_rate": 4.139333691969071e-06, "loss": 0.7418, "step": 720 }, { "epoch": 2.1879563774300617, "grad_norm": 2.1171600818634033, "learning_rate": 4.1107047094188946e-06, "loss": 0.7222, "step": 721 }, { "epoch": 2.190990990990991, "grad_norm": 2.3039426803588867, "learning_rate": 4.082149438613514e-06, "loss": 0.731, "step": 722 }, { "epoch": 2.1940256045519204, "grad_norm": 2.207404375076294, "learning_rate": 4.053668236957135e-06, "loss": 0.7499, "step": 723 }, { "epoch": 2.1970602181128496, "grad_norm": 2.1732053756713867, "learning_rate": 4.025261460926877e-06, "loss": 0.7617, "step": 724 }, { "epoch": 2.200094831673779, "grad_norm": 2.537623167037964, "learning_rate": 3.996929466068344e-06, "loss": 0.72, "step": 725 }, { "epoch": 2.2031294452347083, "grad_norm": 2.2418384552001953, "learning_rate": 3.96867260699116e-06, "loss": 0.727, "step": 726 }, { "epoch": 2.206164058795638, "grad_norm": 2.189655065536499, "learning_rate": 3.940491237364519e-06, "loss": 0.7321, "step": 727 }, { "epoch": 2.209198672356567, "grad_norm": 2.1620960235595703, "learning_rate": 3.912385709912794e-06, "loss": 0.7421, "step": 728 }, { "epoch": 2.2122332859174962, "grad_norm": 2.216374397277832, "learning_rate": 3.884356376411089e-06, "loss": 0.7632, "step": 729 }, { "epoch": 2.215267899478426, "grad_norm": 2.152038335800171, "learning_rate": 3.8564035876808405e-06, "loss": 0.7349, "step": 730 }, { "epoch": 2.218302513039355, "grad_norm": 2.1349170207977295, "learning_rate": 3.828527693585451e-06, "loss": 0.7373, "step": 731 }, { "epoch": 2.2213371266002846, "grad_norm": 2.1102511882781982, "learning_rate": 3.8007290430258712e-06, "loss": 0.7278, "step": 732 }, { "epoch": 2.2243717401612138, "grad_norm": 2.124293088912964, "learning_rate": 3.7730079839362755e-06, "loss": 0.7315, "step": 733 }, { "epoch": 2.2274063537221434, "grad_norm": 2.0747625827789307, "learning_rate": 3.7453648632796745e-06, "loss": 0.7391, "step": 734 }, { "epoch": 2.2304409672830725, "grad_norm": 2.1584184169769287, "learning_rate": 3.7178000270435765e-06, "loss": 0.7526, "step": 735 }, { "epoch": 2.2334755808440017, "grad_norm": 2.0907371044158936, "learning_rate": 3.690313820235686e-06, "loss": 0.7176, "step": 736 }, { "epoch": 2.2365101944049313, "grad_norm": 2.100431442260742, "learning_rate": 3.662906586879542e-06, "loss": 0.7619, "step": 737 }, { "epoch": 2.2395448079658604, "grad_norm": 2.1690993309020996, "learning_rate": 3.6355786700102426e-06, "loss": 0.7385, "step": 738 }, { "epoch": 2.24257942152679, "grad_norm": 2.2165980339050293, "learning_rate": 3.6083304116701535e-06, "loss": 0.7577, "step": 739 }, { "epoch": 2.245614035087719, "grad_norm": 2.1825928688049316, "learning_rate": 3.5811621529045927e-06, "loss": 0.7353, "step": 740 }, { "epoch": 2.2486486486486488, "grad_norm": 2.1485071182250977, "learning_rate": 3.5540742337576083e-06, "loss": 0.7534, "step": 741 }, { "epoch": 2.251683262209578, "grad_norm": 2.193574905395508, "learning_rate": 3.5270669932676926e-06, "loss": 0.7395, "step": 742 }, { "epoch": 2.2547178757705075, "grad_norm": 2.152099847793579, "learning_rate": 3.5001407694635326e-06, "loss": 0.7265, "step": 743 }, { "epoch": 2.2577524893314367, "grad_norm": 2.2395544052124023, "learning_rate": 3.4732958993598153e-06, "loss": 0.7482, "step": 744 }, { "epoch": 2.2607871028923663, "grad_norm": 2.209214448928833, "learning_rate": 3.446532718952966e-06, "loss": 0.7366, "step": 745 }, { "epoch": 2.2638217164532954, "grad_norm": 2.282007932662964, "learning_rate": 3.4198515632169703e-06, "loss": 0.7641, "step": 746 }, { "epoch": 2.2668563300142246, "grad_norm": 2.1614651679992676, "learning_rate": 3.3932527660991877e-06, "loss": 0.7502, "step": 747 }, { "epoch": 2.269890943575154, "grad_norm": 2.306121826171875, "learning_rate": 3.3667366605161322e-06, "loss": 0.7464, "step": 748 }, { "epoch": 2.2729255571360834, "grad_norm": 2.2220146656036377, "learning_rate": 3.340303578349361e-06, "loss": 0.7497, "step": 749 }, { "epoch": 2.275960170697013, "grad_norm": 2.1184372901916504, "learning_rate": 3.313953850441266e-06, "loss": 0.7434, "step": 750 }, { "epoch": 2.275960170697013, "eval_loss": NaN, "eval_runtime": 203.9554, "eval_samples_per_second": 20.387, "eval_steps_per_second": 5.099, "step": 750 }, { "epoch": 2.278994784257942, "grad_norm": 2.2895419597625732, "learning_rate": 3.2876878065909714e-06, "loss": 0.7362, "step": 751 }, { "epoch": 2.2820293978188717, "grad_norm": 2.148529291152954, "learning_rate": 3.2615057755502e-06, "loss": 0.7558, "step": 752 }, { "epoch": 2.285064011379801, "grad_norm": 2.1223018169403076, "learning_rate": 3.2354080850191328e-06, "loss": 0.7591, "step": 753 }, { "epoch": 2.28809862494073, "grad_norm": 2.2450056076049805, "learning_rate": 3.2093950616423397e-06, "loss": 0.7335, "step": 754 }, { "epoch": 2.2911332385016596, "grad_norm": 2.173351526260376, "learning_rate": 3.1834670310046735e-06, "loss": 0.7546, "step": 755 }, { "epoch": 2.2941678520625888, "grad_norm": 2.0341904163360596, "learning_rate": 3.157624317627195e-06, "loss": 0.7447, "step": 756 }, { "epoch": 2.2972024656235184, "grad_norm": 2.0900380611419678, "learning_rate": 3.1318672449631283e-06, "loss": 0.7544, "step": 757 }, { "epoch": 2.3002370791844475, "grad_norm": 2.2536418437957764, "learning_rate": 3.106196135393782e-06, "loss": 0.7326, "step": 758 }, { "epoch": 2.303271692745377, "grad_norm": 2.1906511783599854, "learning_rate": 3.0806113102245395e-06, "loss": 0.7522, "step": 759 }, { "epoch": 2.3063063063063063, "grad_norm": 2.1476638317108154, "learning_rate": 3.055113089680829e-06, "loss": 0.7423, "step": 760 }, { "epoch": 2.3093409198672354, "grad_norm": 2.1808698177337646, "learning_rate": 3.029701792904117e-06, "loss": 0.7612, "step": 761 }, { "epoch": 2.312375533428165, "grad_norm": 2.085947036743164, "learning_rate": 3.00437773794791e-06, "loss": 0.7439, "step": 762 }, { "epoch": 2.315410146989094, "grad_norm": 2.1122984886169434, "learning_rate": 2.979141241773775e-06, "loss": 0.7383, "step": 763 }, { "epoch": 2.318444760550024, "grad_norm": 2.1743266582489014, "learning_rate": 2.953992620247379e-06, "loss": 0.7458, "step": 764 }, { "epoch": 2.321479374110953, "grad_norm": 2.1469690799713135, "learning_rate": 2.9289321881345257e-06, "loss": 0.756, "step": 765 }, { "epoch": 2.3245139876718826, "grad_norm": 2.077439069747925, "learning_rate": 2.9039602590972203e-06, "loss": 0.7316, "step": 766 }, { "epoch": 2.3275486012328117, "grad_norm": 2.4407780170440674, "learning_rate": 2.879077145689746e-06, "loss": 0.7343, "step": 767 }, { "epoch": 2.330583214793741, "grad_norm": 2.1479241847991943, "learning_rate": 2.8542831593547483e-06, "loss": 0.7407, "step": 768 }, { "epoch": 2.3336178283546705, "grad_norm": 2.2073941230773926, "learning_rate": 2.829578610419337e-06, "loss": 0.7471, "step": 769 }, { "epoch": 2.3366524419155996, "grad_norm": 2.0292317867279053, "learning_rate": 2.8049638080912045e-06, "loss": 0.7332, "step": 770 }, { "epoch": 2.339687055476529, "grad_norm": 2.2681589126586914, "learning_rate": 2.780439060454756e-06, "loss": 0.7436, "step": 771 }, { "epoch": 2.3427216690374584, "grad_norm": 2.057676076889038, "learning_rate": 2.75600467446725e-06, "loss": 0.7352, "step": 772 }, { "epoch": 2.345756282598388, "grad_norm": 2.2097206115722656, "learning_rate": 2.7316609559549568e-06, "loss": 0.7275, "step": 773 }, { "epoch": 2.348790896159317, "grad_norm": 2.3039231300354004, "learning_rate": 2.707408209609339e-06, "loss": 0.7556, "step": 774 }, { "epoch": 2.3518255097202467, "grad_norm": 2.097167730331421, "learning_rate": 2.683246738983217e-06, "loss": 0.7404, "step": 775 }, { "epoch": 2.354860123281176, "grad_norm": 2.083677053451538, "learning_rate": 2.6591768464870016e-06, "loss": 0.7359, "step": 776 }, { "epoch": 2.3578947368421055, "grad_norm": 2.0094263553619385, "learning_rate": 2.6351988333848787e-06, "loss": 0.7297, "step": 777 }, { "epoch": 2.3609293504030346, "grad_norm": 2.13173770904541, "learning_rate": 2.611312999791055e-06, "loss": 0.7315, "step": 778 }, { "epoch": 2.363963963963964, "grad_norm": 2.1343095302581787, "learning_rate": 2.587519644666001e-06, "loss": 0.7309, "step": 779 }, { "epoch": 2.3669985775248934, "grad_norm": 2.122774600982666, "learning_rate": 2.5638190658126937e-06, "loss": 0.7428, "step": 780 }, { "epoch": 2.3700331910858226, "grad_norm": 2.1483750343322754, "learning_rate": 2.5402115598729182e-06, "loss": 0.7386, "step": 781 }, { "epoch": 2.373067804646752, "grad_norm": 2.189150810241699, "learning_rate": 2.5166974223235295e-06, "loss": 0.7237, "step": 782 }, { "epoch": 2.3761024182076813, "grad_norm": 2.156003475189209, "learning_rate": 2.493276947472756e-06, "loss": 0.7614, "step": 783 }, { "epoch": 2.379137031768611, "grad_norm": 2.100486993789673, "learning_rate": 2.4699504284565413e-06, "loss": 0.7543, "step": 784 }, { "epoch": 2.38217164532954, "grad_norm": 2.1420738697052, "learning_rate": 2.446718157234832e-06, "loss": 0.7475, "step": 785 }, { "epoch": 2.385206258890469, "grad_norm": 2.205798625946045, "learning_rate": 2.4235804245879723e-06, "loss": 0.7362, "step": 786 }, { "epoch": 2.388240872451399, "grad_norm": 2.193894147872925, "learning_rate": 2.4005375201130275e-06, "loss": 0.7456, "step": 787 }, { "epoch": 2.391275486012328, "grad_norm": 2.1630735397338867, "learning_rate": 2.3775897322201667e-06, "loss": 0.7553, "step": 788 }, { "epoch": 2.3943100995732576, "grad_norm": 2.247066020965576, "learning_rate": 2.354737348129077e-06, "loss": 0.7324, "step": 789 }, { "epoch": 2.3973447131341867, "grad_norm": 2.1332616806030273, "learning_rate": 2.33198065386533e-06, "loss": 0.7413, "step": 790 }, { "epoch": 2.4003793266951163, "grad_norm": 2.1438822746276855, "learning_rate": 2.3093199342568316e-06, "loss": 0.7215, "step": 791 }, { "epoch": 2.4034139402560455, "grad_norm": 2.1765286922454834, "learning_rate": 2.2867554729302545e-06, "loss": 0.7362, "step": 792 }, { "epoch": 2.4064485538169746, "grad_norm": 2.059136390686035, "learning_rate": 2.2642875523074613e-06, "loss": 0.7373, "step": 793 }, { "epoch": 2.4094831673779042, "grad_norm": 2.195065975189209, "learning_rate": 2.2419164536020112e-06, "loss": 0.7363, "step": 794 }, { "epoch": 2.4125177809388334, "grad_norm": 2.2990760803222656, "learning_rate": 2.2196424568156073e-06, "loss": 0.7348, "step": 795 }, { "epoch": 2.415552394499763, "grad_norm": 2.251936435699463, "learning_rate": 2.197465840734596e-06, "loss": 0.7257, "step": 796 }, { "epoch": 2.418587008060692, "grad_norm": 2.142132520675659, "learning_rate": 2.1753868829265046e-06, "loss": 0.7526, "step": 797 }, { "epoch": 2.4216216216216218, "grad_norm": 2.235971450805664, "learning_rate": 2.1534058597365284e-06, "loss": 0.7402, "step": 798 }, { "epoch": 2.424656235182551, "grad_norm": 2.1040003299713135, "learning_rate": 2.1315230462840985e-06, "loss": 0.7271, "step": 799 }, { "epoch": 2.4276908487434805, "grad_norm": 2.153122663497925, "learning_rate": 2.1097387164594406e-06, "loss": 0.7274, "step": 800 }, { "epoch": 2.4276908487434805, "eval_loss": NaN, "eval_runtime": 204.5921, "eval_samples_per_second": 20.323, "eval_steps_per_second": 5.083, "step": 800 }, { "epoch": 2.4307254623044097, "grad_norm": 2.2538816928863525, "learning_rate": 2.0880531429201146e-06, "loss": 0.749, "step": 801 }, { "epoch": 2.4337600758653393, "grad_norm": 2.194967031478882, "learning_rate": 2.0664665970876496e-06, "loss": 0.7395, "step": 802 }, { "epoch": 2.4367946894262684, "grad_norm": 2.1779520511627197, "learning_rate": 2.0449793491441026e-06, "loss": 0.7476, "step": 803 }, { "epoch": 2.4398293029871976, "grad_norm": 2.243781805038452, "learning_rate": 2.0235916680287015e-06, "loss": 0.7268, "step": 804 }, { "epoch": 2.442863916548127, "grad_norm": 2.1497793197631836, "learning_rate": 2.0023038214344827e-06, "loss": 0.76, "step": 805 }, { "epoch": 2.4458985301090563, "grad_norm": 2.209360122680664, "learning_rate": 1.9811160758049163e-06, "loss": 0.7133, "step": 806 }, { "epoch": 2.448933143669986, "grad_norm": 2.1285481452941895, "learning_rate": 1.960028696330596e-06, "loss": 0.7349, "step": 807 }, { "epoch": 2.451967757230915, "grad_norm": 2.16249680519104, "learning_rate": 1.9390419469459066e-06, "loss": 0.7392, "step": 808 }, { "epoch": 2.4550023707918447, "grad_norm": 2.1789779663085938, "learning_rate": 1.9181560903257234e-06, "loss": 0.7299, "step": 809 }, { "epoch": 2.458036984352774, "grad_norm": 2.196904420852661, "learning_rate": 1.8973713878821343e-06, "loss": 0.74, "step": 810 }, { "epoch": 2.461071597913703, "grad_norm": 2.087092399597168, "learning_rate": 1.8766880997611424e-06, "loss": 0.7503, "step": 811 }, { "epoch": 2.4641062114746326, "grad_norm": 2.023298978805542, "learning_rate": 1.8561064848394384e-06, "loss": 0.7409, "step": 812 }, { "epoch": 2.4671408250355618, "grad_norm": 2.2470693588256836, "learning_rate": 1.8356268007211442e-06, "loss": 0.7553, "step": 813 }, { "epoch": 2.4701754385964914, "grad_norm": 2.0922911167144775, "learning_rate": 1.8152493037345942e-06, "loss": 0.7257, "step": 814 }, { "epoch": 2.4732100521574205, "grad_norm": 2.100867986679077, "learning_rate": 1.7949742489291256e-06, "loss": 0.7275, "step": 815 }, { "epoch": 2.47624466571835, "grad_norm": 2.309476375579834, "learning_rate": 1.7748018900718856e-06, "loss": 0.7349, "step": 816 }, { "epoch": 2.4792792792792793, "grad_norm": 2.1686556339263916, "learning_rate": 1.7547324796446553e-06, "loss": 0.746, "step": 817 }, { "epoch": 2.4823138928402084, "grad_norm": 2.161126136779785, "learning_rate": 1.7347662688406907e-06, "loss": 0.7637, "step": 818 }, { "epoch": 2.485348506401138, "grad_norm": 2.0931811332702637, "learning_rate": 1.7149035075615795e-06, "loss": 0.7417, "step": 819 }, { "epoch": 2.488383119962067, "grad_norm": 2.220621347427368, "learning_rate": 1.6951444444141084e-06, "loss": 0.7426, "step": 820 }, { "epoch": 2.491417733522997, "grad_norm": 2.1453609466552734, "learning_rate": 1.6754893267071593e-06, "loss": 0.7277, "step": 821 }, { "epoch": 2.494452347083926, "grad_norm": 2.204213857650757, "learning_rate": 1.6559384004486057e-06, "loss": 0.7456, "step": 822 }, { "epoch": 2.4974869606448555, "grad_norm": 2.1252691745758057, "learning_rate": 1.6364919103422394e-06, "loss": 0.7423, "step": 823 }, { "epoch": 2.5005215742057847, "grad_norm": 2.285757064819336, "learning_rate": 1.6171500997847056e-06, "loss": 0.7377, "step": 824 }, { "epoch": 2.503556187766714, "grad_norm": 2.1519298553466797, "learning_rate": 1.5979132108624572e-06, "loss": 0.7397, "step": 825 }, { "epoch": 2.5065908013276434, "grad_norm": 2.1684153079986572, "learning_rate": 1.5787814843487226e-06, "loss": 0.7355, "step": 826 }, { "epoch": 2.509625414888573, "grad_norm": 2.1787893772125244, "learning_rate": 1.5597551597004968e-06, "loss": 0.7356, "step": 827 }, { "epoch": 2.512660028449502, "grad_norm": 2.22249436378479, "learning_rate": 1.5408344750555382e-06, "loss": 0.7439, "step": 828 }, { "epoch": 2.5156946420104314, "grad_norm": 2.2051734924316406, "learning_rate": 1.522019667229393e-06, "loss": 0.7271, "step": 829 }, { "epoch": 2.518729255571361, "grad_norm": 2.0986456871032715, "learning_rate": 1.5033109717124284e-06, "loss": 0.7385, "step": 830 }, { "epoch": 2.52176386913229, "grad_norm": 2.115734100341797, "learning_rate": 1.4847086226668871e-06, "loss": 0.7317, "step": 831 }, { "epoch": 2.5247984826932193, "grad_norm": 2.02058482170105, "learning_rate": 1.4662128529239572e-06, "loss": 0.7524, "step": 832 }, { "epoch": 2.527833096254149, "grad_norm": 2.1982414722442627, "learning_rate": 1.4478238939808454e-06, "loss": 0.7409, "step": 833 }, { "epoch": 2.5308677098150785, "grad_norm": 2.1642115116119385, "learning_rate": 1.4295419759979079e-06, "loss": 0.7365, "step": 834 }, { "epoch": 2.5339023233760076, "grad_norm": 2.118788003921509, "learning_rate": 1.4113673277957395e-06, "loss": 0.7276, "step": 835 }, { "epoch": 2.536936936936937, "grad_norm": 2.0929243564605713, "learning_rate": 1.393300176852327e-06, "loss": 0.7449, "step": 836 }, { "epoch": 2.5399715504978664, "grad_norm": 2.1339359283447266, "learning_rate": 1.3753407493001968e-06, "loss": 0.7241, "step": 837 }, { "epoch": 2.5430061640587955, "grad_norm": 2.15981125831604, "learning_rate": 1.3574892699235798e-06, "loss": 0.7373, "step": 838 }, { "epoch": 2.546040777619725, "grad_norm": 2.184105396270752, "learning_rate": 1.339745962155613e-06, "loss": 0.7283, "step": 839 }, { "epoch": 2.5490753911806543, "grad_norm": 2.0236942768096924, "learning_rate": 1.3221110480755306e-06, "loss": 0.7279, "step": 840 }, { "epoch": 2.552110004741584, "grad_norm": 2.0451905727386475, "learning_rate": 1.3045847484058748e-06, "loss": 0.7417, "step": 841 }, { "epoch": 2.555144618302513, "grad_norm": 2.113330602645874, "learning_rate": 1.287167282509767e-06, "loss": 0.7604, "step": 842 }, { "epoch": 2.558179231863442, "grad_norm": 2.0830495357513428, "learning_rate": 1.2698588683881185e-06, "loss": 0.7579, "step": 843 }, { "epoch": 2.561213845424372, "grad_norm": 2.065068483352661, "learning_rate": 1.252659722676941e-06, "loss": 0.7268, "step": 844 }, { "epoch": 2.564248458985301, "grad_norm": 2.157485008239746, "learning_rate": 1.2355700606446119e-06, "loss": 0.7305, "step": 845 }, { "epoch": 2.5672830725462306, "grad_norm": 2.135056257247925, "learning_rate": 1.2185900961891794e-06, "loss": 0.7332, "step": 846 }, { "epoch": 2.5703176861071597, "grad_norm": 2.1909451484680176, "learning_rate": 1.2017200418357077e-06, "loss": 0.7402, "step": 847 }, { "epoch": 2.5733522996680893, "grad_norm": 2.2838058471679688, "learning_rate": 1.184960108733586e-06, "loss": 0.7336, "step": 848 }, { "epoch": 2.5763869132290185, "grad_norm": 2.0348660945892334, "learning_rate": 1.1683105066539068e-06, "loss": 0.7367, "step": 849 }, { "epoch": 2.5794215267899476, "grad_norm": 2.088353157043457, "learning_rate": 1.151771443986842e-06, "loss": 0.749, "step": 850 }, { "epoch": 2.5794215267899476, "eval_loss": NaN, "eval_runtime": 205.2076, "eval_samples_per_second": 20.262, "eval_steps_per_second": 5.068, "step": 850 }, { "epoch": 2.5824561403508772, "grad_norm": 2.1730659008026123, "learning_rate": 1.1353431277390125e-06, "loss": 0.7359, "step": 851 }, { "epoch": 2.5854907539118064, "grad_norm": 2.10697078704834, "learning_rate": 1.1190257635309276e-06, "loss": 0.7223, "step": 852 }, { "epoch": 2.588525367472736, "grad_norm": 2.1502668857574463, "learning_rate": 1.1028195555943877e-06, "loss": 0.7265, "step": 853 }, { "epoch": 2.591559981033665, "grad_norm": 2.0735018253326416, "learning_rate": 1.0867247067699315e-06, "loss": 0.7345, "step": 854 }, { "epoch": 2.5945945945945947, "grad_norm": 2.1325292587280273, "learning_rate": 1.0707414185043163e-06, "loss": 0.746, "step": 855 }, { "epoch": 2.597629208155524, "grad_norm": 2.1418328285217285, "learning_rate": 1.0548698908479671e-06, "loss": 0.7401, "step": 856 }, { "epoch": 2.600663821716453, "grad_norm": 2.061847448348999, "learning_rate": 1.0391103224524957e-06, "loss": 0.7319, "step": 857 }, { "epoch": 2.6036984352773827, "grad_norm": 2.1931376457214355, "learning_rate": 1.0234629105682104e-06, "loss": 0.7429, "step": 858 }, { "epoch": 2.6067330488383122, "grad_norm": 2.1396734714508057, "learning_rate": 1.0079278510416313e-06, "loss": 0.7369, "step": 859 }, { "epoch": 2.6097676623992414, "grad_norm": 2.2338297367095947, "learning_rate": 9.925053383130667e-07, "loss": 0.745, "step": 860 }, { "epoch": 2.6128022759601706, "grad_norm": 2.120897054672241, "learning_rate": 9.771955654141496e-07, "loss": 0.7444, "step": 861 }, { "epoch": 2.6158368895211, "grad_norm": 2.204875946044922, "learning_rate": 9.619987239654405e-07, "loss": 0.7293, "step": 862 }, { "epoch": 2.6188715030820293, "grad_norm": 2.228181838989258, "learning_rate": 9.469150041740338e-07, "loss": 0.7251, "step": 863 }, { "epoch": 2.6219061166429585, "grad_norm": 2.170053720474243, "learning_rate": 9.319445948311534e-07, "loss": 0.7444, "step": 864 }, { "epoch": 2.624940730203888, "grad_norm": 2.163527011871338, "learning_rate": 9.170876833098119e-07, "loss": 0.7568, "step": 865 }, { "epoch": 2.6279753437648177, "grad_norm": 2.1692657470703125, "learning_rate": 9.023444555624572e-07, "loss": 0.736, "step": 866 }, { "epoch": 2.631009957325747, "grad_norm": 2.0737569332122803, "learning_rate": 8.87715096118642e-07, "loss": 0.7368, "step": 867 }, { "epoch": 2.634044570886676, "grad_norm": 2.05617094039917, "learning_rate": 8.731997880827258e-07, "loss": 0.7334, "step": 868 }, { "epoch": 2.6370791844476056, "grad_norm": 1.9645661115646362, "learning_rate": 8.587987131315656e-07, "loss": 0.7422, "step": 869 }, { "epoch": 2.6401137980085347, "grad_norm": 2.0881567001342773, "learning_rate": 8.445120515122552e-07, "loss": 0.7296, "step": 870 }, { "epoch": 2.6431484115694643, "grad_norm": 2.2319769859313965, "learning_rate": 8.303399820398672e-07, "loss": 0.7502, "step": 871 }, { "epoch": 2.6461830251303935, "grad_norm": 2.1117403507232666, "learning_rate": 8.162826820952097e-07, "loss": 0.7319, "step": 872 }, { "epoch": 2.649217638691323, "grad_norm": 2.056861400604248, "learning_rate": 8.023403276226127e-07, "loss": 0.7429, "step": 873 }, { "epoch": 2.6522522522522523, "grad_norm": 2.140435218811035, "learning_rate": 7.885130931277219e-07, "loss": 0.7418, "step": 874 }, { "epoch": 2.6552868658131814, "grad_norm": 2.11114239692688, "learning_rate": 7.74801151675314e-07, "loss": 0.7521, "step": 875 }, { "epoch": 2.658321479374111, "grad_norm": 2.0694892406463623, "learning_rate": 7.612046748871327e-07, "loss": 0.7385, "step": 876 }, { "epoch": 2.66135609293504, "grad_norm": 2.155579090118408, "learning_rate": 7.477238329397419e-07, "loss": 0.7427, "step": 877 }, { "epoch": 2.6643907064959698, "grad_norm": 2.223646402359009, "learning_rate": 7.343587945623908e-07, "loss": 0.731, "step": 878 }, { "epoch": 2.667425320056899, "grad_norm": 2.047801971435547, "learning_rate": 7.211097270349065e-07, "loss": 0.7426, "step": 879 }, { "epoch": 2.6704599336178285, "grad_norm": 2.119241714477539, "learning_rate": 7.07976796185601e-07, "loss": 0.7407, "step": 880 }, { "epoch": 2.6734945471787577, "grad_norm": 2.2282469272613525, "learning_rate": 6.949601663891891e-07, "loss": 0.7541, "step": 881 }, { "epoch": 2.676529160739687, "grad_norm": 2.175909996032715, "learning_rate": 6.820600005647382e-07, "loss": 0.7461, "step": 882 }, { "epoch": 2.6795637743006164, "grad_norm": 2.047321081161499, "learning_rate": 6.692764601736268e-07, "loss": 0.7197, "step": 883 }, { "epoch": 2.682598387861546, "grad_norm": 2.0702617168426514, "learning_rate": 6.566097052175213e-07, "loss": 0.7656, "step": 884 }, { "epoch": 2.685633001422475, "grad_norm": 2.1683459281921387, "learning_rate": 6.440598942363796e-07, "loss": 0.7521, "step": 885 }, { "epoch": 2.6886676149834043, "grad_norm": 2.067746162414551, "learning_rate": 6.316271843064536e-07, "loss": 0.7531, "step": 886 }, { "epoch": 2.691702228544334, "grad_norm": 2.1310875415802, "learning_rate": 6.193117310383412e-07, "loss": 0.7538, "step": 887 }, { "epoch": 2.694736842105263, "grad_norm": 2.1403868198394775, "learning_rate": 6.071136885750272e-07, "loss": 0.7373, "step": 888 }, { "epoch": 2.6977714556661923, "grad_norm": 2.049807071685791, "learning_rate": 5.950332095899547e-07, "loss": 0.738, "step": 889 }, { "epoch": 2.700806069227122, "grad_norm": 2.080238103866577, "learning_rate": 5.830704452851166e-07, "loss": 0.7212, "step": 890 }, { "epoch": 2.7038406827880515, "grad_norm": 2.1021158695220947, "learning_rate": 5.71225545389158e-07, "loss": 0.7411, "step": 891 }, { "epoch": 2.7068752963489806, "grad_norm": 2.0948758125305176, "learning_rate": 5.594986581555173e-07, "loss": 0.7369, "step": 892 }, { "epoch": 2.7099099099099098, "grad_norm": 2.1047518253326416, "learning_rate": 5.478899303605512e-07, "loss": 0.7289, "step": 893 }, { "epoch": 2.7129445234708394, "grad_norm": 2.1207687854766846, "learning_rate": 5.363995073017047e-07, "loss": 0.7385, "step": 894 }, { "epoch": 2.7159791370317685, "grad_norm": 2.051896572113037, "learning_rate": 5.250275327957033e-07, "loss": 0.7244, "step": 895 }, { "epoch": 2.719013750592698, "grad_norm": 2.220273017883301, "learning_rate": 5.137741491767345e-07, "loss": 0.7522, "step": 896 }, { "epoch": 2.7220483641536273, "grad_norm": 2.1100893020629883, "learning_rate": 5.026394972946813e-07, "loss": 0.7276, "step": 897 }, { "epoch": 2.725082977714557, "grad_norm": 2.300666332244873, "learning_rate": 4.91623716513352e-07, "loss": 0.7404, "step": 898 }, { "epoch": 2.728117591275486, "grad_norm": 2.2343952655792236, "learning_rate": 4.807269447087348e-07, "loss": 0.7474, "step": 899 }, { "epoch": 2.731152204836415, "grad_norm": 2.1582469940185547, "learning_rate": 4.6994931826728094e-07, "loss": 0.7275, "step": 900 }, { "epoch": 2.731152204836415, "eval_loss": NaN, "eval_runtime": 204.8855, "eval_samples_per_second": 20.294, "eval_steps_per_second": 5.076, "step": 900 }, { "epoch": 2.734186818397345, "grad_norm": 2.057573080062866, "learning_rate": 4.592909720841843e-07, "loss": 0.7351, "step": 901 }, { "epoch": 2.737221431958274, "grad_norm": 2.320993423461914, "learning_rate": 4.487520395617029e-07, "loss": 0.7086, "step": 902 }, { "epoch": 2.7402560455192035, "grad_norm": 2.2511098384857178, "learning_rate": 4.3833265260749157e-07, "loss": 0.723, "step": 903 }, { "epoch": 2.7432906590801327, "grad_norm": 2.182513952255249, "learning_rate": 4.280329416329365e-07, "loss": 0.7196, "step": 904 }, { "epoch": 2.7463252726410623, "grad_norm": 2.1579484939575195, "learning_rate": 4.178530355515409e-07, "loss": 0.7334, "step": 905 }, { "epoch": 2.7493598862019915, "grad_norm": 2.091196060180664, "learning_rate": 4.077930617773007e-07, "loss": 0.7387, "step": 906 }, { "epoch": 2.7523944997629206, "grad_norm": 2.1611766815185547, "learning_rate": 3.97853146223105e-07, "loss": 0.7515, "step": 907 }, { "epoch": 2.75542911332385, "grad_norm": 2.240387201309204, "learning_rate": 3.880334132991792e-07, "loss": 0.7377, "step": 908 }, { "epoch": 2.7584637268847794, "grad_norm": 2.1546630859375, "learning_rate": 3.783339859115065e-07, "loss": 0.7388, "step": 909 }, { "epoch": 2.761498340445709, "grad_norm": 2.0892395973205566, "learning_rate": 3.687549854603023e-07, "loss": 0.7488, "step": 910 }, { "epoch": 2.764532954006638, "grad_norm": 2.182509422302246, "learning_rate": 3.5929653183849444e-07, "loss": 0.736, "step": 911 }, { "epoch": 2.7675675675675677, "grad_norm": 2.198559522628784, "learning_rate": 3.49958743430211e-07, "loss": 0.736, "step": 912 }, { "epoch": 2.770602181128497, "grad_norm": 2.188133955001831, "learning_rate": 3.4074173710931804e-07, "loss": 0.7345, "step": 913 }, { "epoch": 2.773636794689426, "grad_norm": 2.133892774581909, "learning_rate": 3.3164562823793654e-07, "loss": 0.7275, "step": 914 }, { "epoch": 2.7766714082503556, "grad_norm": 2.159824848175049, "learning_rate": 3.226705306650113e-07, "loss": 0.7198, "step": 915 }, { "epoch": 2.7797060218112852, "grad_norm": 2.1087584495544434, "learning_rate": 3.1381655672488634e-07, "loss": 0.7381, "step": 916 }, { "epoch": 2.7827406353722144, "grad_norm": 2.0807528495788574, "learning_rate": 3.050838172358883e-07, "loss": 0.7386, "step": 917 }, { "epoch": 2.7857752489331435, "grad_norm": 2.1473801136016846, "learning_rate": 2.9647242149895005e-07, "loss": 0.7336, "step": 918 }, { "epoch": 2.788809862494073, "grad_norm": 2.219571352005005, "learning_rate": 2.879824772962381e-07, "loss": 0.7438, "step": 919 }, { "epoch": 2.7918444760550023, "grad_norm": 2.144059419631958, "learning_rate": 2.796140908898026e-07, "loss": 0.7338, "step": 920 }, { "epoch": 2.7948790896159315, "grad_norm": 2.1552734375, "learning_rate": 2.7136736702025436e-07, "loss": 0.7345, "step": 921 }, { "epoch": 2.797913703176861, "grad_norm": 2.18730092048645, "learning_rate": 2.632424089054419e-07, "loss": 0.7222, "step": 922 }, { "epoch": 2.8009483167377907, "grad_norm": 2.233304500579834, "learning_rate": 2.552393182391677e-07, "loss": 0.7474, "step": 923 }, { "epoch": 2.80398293029872, "grad_norm": 2.041757583618164, "learning_rate": 2.473581951899184e-07, "loss": 0.7329, "step": 924 }, { "epoch": 2.807017543859649, "grad_norm": 2.1639747619628906, "learning_rate": 2.395991383995999e-07, "loss": 0.748, "step": 925 }, { "epoch": 2.8100521574205786, "grad_norm": 2.0612905025482178, "learning_rate": 2.3196224498231447e-07, "loss": 0.7311, "step": 926 }, { "epoch": 2.8130867709815077, "grad_norm": 2.260965347290039, "learning_rate": 2.2444761052313857e-07, "loss": 0.7353, "step": 927 }, { "epoch": 2.8161213845424373, "grad_norm": 2.1008665561676025, "learning_rate": 2.1705532907692617e-07, "loss": 0.7677, "step": 928 }, { "epoch": 2.8191559981033665, "grad_norm": 2.20302414894104, "learning_rate": 2.0978549316713615e-07, "loss": 0.7447, "step": 929 }, { "epoch": 2.822190611664296, "grad_norm": 2.100362539291382, "learning_rate": 2.0263819378466888e-07, "loss": 0.7393, "step": 930 }, { "epoch": 2.8252252252252252, "grad_norm": 2.2228076457977295, "learning_rate": 1.9561352038673264e-07, "loss": 0.7501, "step": 931 }, { "epoch": 2.8282598387861544, "grad_norm": 2.2403063774108887, "learning_rate": 1.8871156089572018e-07, "loss": 0.7508, "step": 932 }, { "epoch": 2.831294452347084, "grad_norm": 2.1292290687561035, "learning_rate": 1.8193240169810943e-07, "loss": 0.7362, "step": 933 }, { "epoch": 2.834329065908013, "grad_norm": 2.029907703399658, "learning_rate": 1.752761276433801e-07, "loss": 0.7504, "step": 934 }, { "epoch": 2.8373636794689427, "grad_norm": 2.175046682357788, "learning_rate": 1.6874282204295765e-07, "loss": 0.7437, "step": 935 }, { "epoch": 2.840398293029872, "grad_norm": 2.150054693222046, "learning_rate": 1.623325666691644e-07, "loss": 0.7388, "step": 936 }, { "epoch": 2.8434329065908015, "grad_norm": 2.2034387588500977, "learning_rate": 1.5604544175419901e-07, "loss": 0.7451, "step": 937 }, { "epoch": 2.8464675201517307, "grad_norm": 2.1042325496673584, "learning_rate": 1.4988152598913063e-07, "loss": 0.7535, "step": 938 }, { "epoch": 2.84950213371266, "grad_norm": 2.1048696041107178, "learning_rate": 1.4384089652291544e-07, "loss": 0.745, "step": 939 }, { "epoch": 2.8525367472735894, "grad_norm": 2.3105156421661377, "learning_rate": 1.3792362896143164e-07, "loss": 0.7233, "step": 940 }, { "epoch": 2.8555713608345186, "grad_norm": 2.1417715549468994, "learning_rate": 1.3212979736653142e-07, "loss": 0.7467, "step": 941 }, { "epoch": 2.858605974395448, "grad_norm": 2.1809537410736084, "learning_rate": 1.2645947425511397e-07, "loss": 0.7432, "step": 942 }, { "epoch": 2.8616405879563773, "grad_norm": 2.2112855911254883, "learning_rate": 1.209127305982205e-07, "loss": 0.7487, "step": 943 }, { "epoch": 2.864675201517307, "grad_norm": 2.059002161026001, "learning_rate": 1.1548963582013961e-07, "loss": 0.734, "step": 944 }, { "epoch": 2.867709815078236, "grad_norm": 2.04377818107605, "learning_rate": 1.1019025779754666e-07, "loss": 0.7213, "step": 945 }, { "epoch": 2.8707444286391652, "grad_norm": 2.0912530422210693, "learning_rate": 1.0501466285865124e-07, "loss": 0.7548, "step": 946 }, { "epoch": 2.873779042200095, "grad_norm": 2.2096190452575684, "learning_rate": 9.996291578236228e-08, "loss": 0.7246, "step": 947 }, { "epoch": 2.8768136557610244, "grad_norm": 2.284653663635254, "learning_rate": 9.503507979748305e-08, "loss": 0.7339, "step": 948 }, { "epoch": 2.8798482693219536, "grad_norm": 2.097752332687378, "learning_rate": 9.023121658191636e-08, "loss": 0.7303, "step": 949 }, { "epoch": 2.8828828828828827, "grad_norm": 2.171391010284424, "learning_rate": 8.555138626189619e-08, "loss": 0.7503, "step": 950 }, { "epoch": 2.8828828828828827, "eval_loss": NaN, "eval_runtime": 204.3654, "eval_samples_per_second": 20.346, "eval_steps_per_second": 5.089, "step": 950 }, { "epoch": 2.8859174964438123, "grad_norm": 2.1366286277770996, "learning_rate": 8.099564741123167e-08, "loss": 0.7241, "step": 951 }, { "epoch": 2.8889521100047415, "grad_norm": 2.0811874866485596, "learning_rate": 7.656405705057434e-08, "loss": 0.7504, "step": 952 }, { "epoch": 2.8919867235656707, "grad_norm": 2.080226421356201, "learning_rate": 7.225667064670761e-08, "loss": 0.7355, "step": 953 }, { "epoch": 2.8950213371266003, "grad_norm": 2.200331211090088, "learning_rate": 6.807354211184613e-08, "loss": 0.7427, "step": 954 }, { "epoch": 2.89805595068753, "grad_norm": 2.038591146469116, "learning_rate": 6.401472380297091e-08, "loss": 0.7379, "step": 955 }, { "epoch": 2.901090564248459, "grad_norm": 2.1141886711120605, "learning_rate": 6.008026652116306e-08, "loss": 0.7376, "step": 956 }, { "epoch": 2.904125177809388, "grad_norm": 2.0803427696228027, "learning_rate": 5.6270219510975445e-08, "loss": 0.7424, "step": 957 }, { "epoch": 2.9071597913703178, "grad_norm": 2.0900638103485107, "learning_rate": 5.258463045981432e-08, "loss": 0.7243, "step": 958 }, { "epoch": 2.910194404931247, "grad_norm": 2.133312463760376, "learning_rate": 4.902354549733979e-08, "loss": 0.7257, "step": 959 }, { "epoch": 2.9132290184921765, "grad_norm": 2.2082343101501465, "learning_rate": 4.5587009194894005e-08, "loss": 0.7346, "step": 960 }, { "epoch": 2.9162636320531057, "grad_norm": 2.0872416496276855, "learning_rate": 4.227506456493835e-08, "loss": 0.7437, "step": 961 }, { "epoch": 2.9192982456140353, "grad_norm": 2.136866807937622, "learning_rate": 3.908775306051604e-08, "loss": 0.7362, "step": 962 }, { "epoch": 2.9223328591749644, "grad_norm": 2.125523805618286, "learning_rate": 3.602511457473479e-08, "loss": 0.7298, "step": 963 }, { "epoch": 2.9253674727358936, "grad_norm": 2.055690050125122, "learning_rate": 3.3087187440268284e-08, "loss": 0.7285, "step": 964 }, { "epoch": 2.928402086296823, "grad_norm": 2.26811146736145, "learning_rate": 3.027400842887218e-08, "loss": 0.732, "step": 965 }, { "epoch": 2.9314366998577523, "grad_norm": 2.0984137058258057, "learning_rate": 2.758561275092886e-08, "loss": 0.7238, "step": 966 }, { "epoch": 2.934471313418682, "grad_norm": 2.117258310317993, "learning_rate": 2.5022034055003363e-08, "loss": 0.7604, "step": 967 }, { "epoch": 2.937505926979611, "grad_norm": 2.135441541671753, "learning_rate": 2.2583304427421516e-08, "loss": 0.7358, "step": 968 }, { "epoch": 2.9405405405405407, "grad_norm": 2.1416280269622803, "learning_rate": 2.0269454391874665e-08, "loss": 0.7436, "step": 969 }, { "epoch": 2.94357515410147, "grad_norm": 2.1733927726745605, "learning_rate": 1.80805129090289e-08, "loss": 0.7308, "step": 970 }, { "epoch": 2.946609767662399, "grad_norm": 2.1229727268218994, "learning_rate": 1.6016507376169776e-08, "loss": 0.7577, "step": 971 }, { "epoch": 2.9496443812233286, "grad_norm": 2.283834218978882, "learning_rate": 1.4077463626852584e-08, "loss": 0.748, "step": 972 }, { "epoch": 2.952678994784258, "grad_norm": 2.205937623977661, "learning_rate": 1.2263405930585947e-08, "loss": 0.7335, "step": 973 }, { "epoch": 2.9557136083451874, "grad_norm": 2.111551284790039, "learning_rate": 1.0574356992525403e-08, "loss": 0.7295, "step": 974 }, { "epoch": 2.9587482219061165, "grad_norm": 2.058469295501709, "learning_rate": 9.010337953185843e-09, "loss": 0.7547, "step": 975 }, { "epoch": 2.961782835467046, "grad_norm": 2.1459245681762695, "learning_rate": 7.571368388181732e-09, "loss": 0.7368, "step": 976 }, { "epoch": 2.9648174490279753, "grad_norm": 2.1419780254364014, "learning_rate": 6.257466307980631e-09, "loss": 0.7364, "step": 977 }, { "epoch": 2.9678520625889044, "grad_norm": 2.1368930339813232, "learning_rate": 5.068648157675604e-09, "loss": 0.7243, "step": 978 }, { "epoch": 2.970886676149834, "grad_norm": 2.072345733642578, "learning_rate": 4.00492881678427e-09, "loss": 0.7229, "step": 979 }, { "epoch": 2.9739212897107636, "grad_norm": 2.173828363418579, "learning_rate": 3.0663215990534013e-09, "loss": 0.711, "step": 980 }, { "epoch": 2.976955903271693, "grad_norm": 2.117293357849121, "learning_rate": 2.2528382523057115e-09, "loss": 0.7648, "step": 981 }, { "epoch": 2.979990516832622, "grad_norm": 2.126967191696167, "learning_rate": 1.564488958279986e-09, "loss": 0.7359, "step": 982 }, { "epoch": 2.9830251303935515, "grad_norm": 2.086338520050049, "learning_rate": 1.0012823325111776e-09, "loss": 0.7575, "step": 983 }, { "epoch": 2.9860597439544807, "grad_norm": 2.0174200534820557, "learning_rate": 5.632254242204926e-10, "loss": 0.7324, "step": 984 }, { "epoch": 2.9890943575154103, "grad_norm": 2.0716025829315186, "learning_rate": 2.503237162254646e-10, "loss": 0.7482, "step": 985 }, { "epoch": 2.9921289710763395, "grad_norm": 2.0842347145080566, "learning_rate": 6.258112487667056e-11, "loss": 0.7334, "step": 986 }, { "epoch": 2.995163584637269, "grad_norm": 2.1061208248138428, "learning_rate": 0.0, "loss": 0.7299, "step": 987 }, { "epoch": 2.995163584637269, "step": 987, "total_flos": 4.415483185831556e+19, "train_loss": 0.7698779804849093, "train_runtime": 131039.1076, "train_samples_per_second": 3.863, "train_steps_per_second": 0.008 } ], "logging_steps": 1.0, "max_steps": 987, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.415483185831556e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }