diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7103 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.995163584637269, + "eval_steps": 50, + "global_step": 987, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0030346135609293505, + "grad_norm": 5.274968147277832, + "learning_rate": 2.0202020202020205e-07, + "loss": 0.8522, + "step": 1 + }, + { + "epoch": 0.006069227121858701, + "grad_norm": 5.754838466644287, + "learning_rate": 4.040404040404041e-07, + "loss": 0.854, + "step": 2 + }, + { + "epoch": 0.009103840682788052, + "grad_norm": 5.632802963256836, + "learning_rate": 6.060606060606061e-07, + "loss": 0.873, + "step": 3 + }, + { + "epoch": 0.012138454243717402, + "grad_norm": 5.394160747528076, + "learning_rate": 8.080808080808082e-07, + "loss": 0.8563, + "step": 4 + }, + { + "epoch": 0.015173067804646752, + "grad_norm": 4.5878424644470215, + "learning_rate": 1.01010101010101e-06, + "loss": 0.8425, + "step": 5 + }, + { + "epoch": 0.018207681365576104, + "grad_norm": 4.280163764953613, + "learning_rate": 1.2121212121212122e-06, + "loss": 0.8569, + "step": 6 + }, + { + "epoch": 0.021242294926505454, + "grad_norm": 3.5766515731811523, + "learning_rate": 1.4141414141414143e-06, + "loss": 0.8309, + "step": 7 + }, + { + "epoch": 0.024276908487434804, + "grad_norm": 4.070709228515625, + "learning_rate": 1.6161616161616164e-06, + "loss": 0.825, + "step": 8 + }, + { + "epoch": 0.027311522048364154, + "grad_norm": 4.395944595336914, + "learning_rate": 1.8181818181818183e-06, + "loss": 0.8419, + "step": 9 + }, + { + "epoch": 0.030346135609293504, + "grad_norm": 3.765310764312744, + "learning_rate": 2.02020202020202e-06, + "loss": 0.8426, + "step": 10 + }, + { + "epoch": 0.03338074917022286, + "grad_norm": 2.906996488571167, + "learning_rate": 2.222222222222222e-06, + "loss": 0.8263, + "step": 11 + }, + { + "epoch": 0.03641536273115221, + "grad_norm": 2.8421216011047363, + "learning_rate": 2.4242424242424244e-06, + "loss": 0.8242, + "step": 12 + }, + { + "epoch": 0.03944997629208156, + "grad_norm": 2.6818017959594727, + "learning_rate": 2.6262626262626267e-06, + "loss": 0.8206, + "step": 13 + }, + { + "epoch": 0.04248458985301091, + "grad_norm": 2.8597841262817383, + "learning_rate": 2.8282828282828286e-06, + "loss": 0.8201, + "step": 14 + }, + { + "epoch": 0.04551920341394026, + "grad_norm": 2.673666000366211, + "learning_rate": 3.0303030303030305e-06, + "loss": 0.8215, + "step": 15 + }, + { + "epoch": 0.04855381697486961, + "grad_norm": 2.636281728744507, + "learning_rate": 3.232323232323233e-06, + "loss": 0.8037, + "step": 16 + }, + { + "epoch": 0.05158843053579896, + "grad_norm": 2.502042770385742, + "learning_rate": 3.4343434343434347e-06, + "loss": 0.8162, + "step": 17 + }, + { + "epoch": 0.05462304409672831, + "grad_norm": 2.562307834625244, + "learning_rate": 3.6363636363636366e-06, + "loss": 0.8066, + "step": 18 + }, + { + "epoch": 0.05765765765765766, + "grad_norm": 2.6667816638946533, + "learning_rate": 3.8383838383838385e-06, + "loss": 0.8197, + "step": 19 + }, + { + "epoch": 0.06069227121858701, + "grad_norm": 2.293628215789795, + "learning_rate": 4.04040404040404e-06, + "loss": 0.8092, + "step": 20 + }, + { + "epoch": 0.06372688477951635, + "grad_norm": 2.4843485355377197, + "learning_rate": 4.242424242424243e-06, + "loss": 0.8082, + "step": 21 + }, + { + "epoch": 0.06676149834044572, + "grad_norm": 2.5178024768829346, + "learning_rate": 4.444444444444444e-06, + "loss": 0.8165, + "step": 22 + }, + { + "epoch": 0.06979611190137507, + "grad_norm": 2.6686830520629883, + "learning_rate": 4.646464646464647e-06, + "loss": 0.773, + "step": 23 + }, + { + "epoch": 0.07283072546230442, + "grad_norm": 2.2187390327453613, + "learning_rate": 4.848484848484849e-06, + "loss": 0.8026, + "step": 24 + }, + { + "epoch": 0.07586533902323377, + "grad_norm": 2.4837453365325928, + "learning_rate": 5.0505050505050515e-06, + "loss": 0.7962, + "step": 25 + }, + { + "epoch": 0.07889995258416312, + "grad_norm": 2.3953895568847656, + "learning_rate": 5.252525252525253e-06, + "loss": 0.7902, + "step": 26 + }, + { + "epoch": 0.08193456614509247, + "grad_norm": 2.384993553161621, + "learning_rate": 5.4545454545454545e-06, + "loss": 0.8133, + "step": 27 + }, + { + "epoch": 0.08496917970602182, + "grad_norm": 2.5588831901550293, + "learning_rate": 5.656565656565657e-06, + "loss": 0.8043, + "step": 28 + }, + { + "epoch": 0.08800379326695117, + "grad_norm": 2.254196882247925, + "learning_rate": 5.858585858585859e-06, + "loss": 0.7967, + "step": 29 + }, + { + "epoch": 0.09103840682788052, + "grad_norm": 2.2686617374420166, + "learning_rate": 6.060606060606061e-06, + "loss": 0.8036, + "step": 30 + }, + { + "epoch": 0.09407302038880987, + "grad_norm": 2.3785650730133057, + "learning_rate": 6.262626262626264e-06, + "loss": 0.799, + "step": 31 + }, + { + "epoch": 0.09710763394973922, + "grad_norm": 2.4035139083862305, + "learning_rate": 6.464646464646466e-06, + "loss": 0.8086, + "step": 32 + }, + { + "epoch": 0.10014224751066857, + "grad_norm": 2.2935521602630615, + "learning_rate": 6.666666666666667e-06, + "loss": 0.7842, + "step": 33 + }, + { + "epoch": 0.10317686107159792, + "grad_norm": 2.3163795471191406, + "learning_rate": 6.868686868686869e-06, + "loss": 0.8041, + "step": 34 + }, + { + "epoch": 0.10621147463252727, + "grad_norm": 2.3953793048858643, + "learning_rate": 7.070707070707071e-06, + "loss": 0.7929, + "step": 35 + }, + { + "epoch": 0.10924608819345662, + "grad_norm": 2.4116768836975098, + "learning_rate": 7.272727272727273e-06, + "loss": 0.7943, + "step": 36 + }, + { + "epoch": 0.11228070175438597, + "grad_norm": 2.5221264362335205, + "learning_rate": 7.474747474747476e-06, + "loss": 0.8036, + "step": 37 + }, + { + "epoch": 0.11531531531531532, + "grad_norm": 2.4730491638183594, + "learning_rate": 7.676767676767677e-06, + "loss": 0.8063, + "step": 38 + }, + { + "epoch": 0.11834992887624467, + "grad_norm": 2.1653268337249756, + "learning_rate": 7.87878787878788e-06, + "loss": 0.8009, + "step": 39 + }, + { + "epoch": 0.12138454243717402, + "grad_norm": 2.266632080078125, + "learning_rate": 8.08080808080808e-06, + "loss": 0.7825, + "step": 40 + }, + { + "epoch": 0.12441915599810337, + "grad_norm": 2.2630982398986816, + "learning_rate": 8.282828282828283e-06, + "loss": 0.8004, + "step": 41 + }, + { + "epoch": 0.1274537695590327, + "grad_norm": 2.403125762939453, + "learning_rate": 8.484848484848486e-06, + "loss": 0.7911, + "step": 42 + }, + { + "epoch": 0.13048838311996205, + "grad_norm": 2.4483842849731445, + "learning_rate": 8.686868686868687e-06, + "loss": 0.7861, + "step": 43 + }, + { + "epoch": 0.13352299668089143, + "grad_norm": 2.346780776977539, + "learning_rate": 8.888888888888888e-06, + "loss": 0.7969, + "step": 44 + }, + { + "epoch": 0.13655761024182078, + "grad_norm": 2.3152859210968018, + "learning_rate": 9.090909090909091e-06, + "loss": 0.7842, + "step": 45 + }, + { + "epoch": 0.13959222380275013, + "grad_norm": 2.2333388328552246, + "learning_rate": 9.292929292929294e-06, + "loss": 0.7886, + "step": 46 + }, + { + "epoch": 0.14262683736367948, + "grad_norm": 2.360898017883301, + "learning_rate": 9.494949494949497e-06, + "loss": 0.8146, + "step": 47 + }, + { + "epoch": 0.14566145092460883, + "grad_norm": 2.5461983680725098, + "learning_rate": 9.696969696969698e-06, + "loss": 0.8005, + "step": 48 + }, + { + "epoch": 0.14869606448553818, + "grad_norm": 2.4684877395629883, + "learning_rate": 9.8989898989899e-06, + "loss": 0.8031, + "step": 49 + }, + { + "epoch": 0.15173067804646753, + "grad_norm": 2.3025074005126953, + "learning_rate": 1.0101010101010103e-05, + "loss": 0.7828, + "step": 50 + }, + { + "epoch": 0.15173067804646753, + "eval_loss": NaN, + "eval_runtime": 233.599, + "eval_samples_per_second": 17.8, + "eval_steps_per_second": 4.452, + "step": 50 + }, + { + "epoch": 0.15476529160739688, + "grad_norm": 2.4276065826416016, + "learning_rate": 1.0303030303030304e-05, + "loss": 0.7801, + "step": 51 + }, + { + "epoch": 0.15779990516832623, + "grad_norm": 2.279287815093994, + "learning_rate": 1.0505050505050507e-05, + "loss": 0.8055, + "step": 52 + }, + { + "epoch": 0.16083451872925558, + "grad_norm": 2.7459404468536377, + "learning_rate": 1.0707070707070708e-05, + "loss": 0.7864, + "step": 53 + }, + { + "epoch": 0.16386913229018493, + "grad_norm": 2.3488147258758545, + "learning_rate": 1.0909090909090909e-05, + "loss": 0.7869, + "step": 54 + }, + { + "epoch": 0.16690374585111428, + "grad_norm": 2.547166109085083, + "learning_rate": 1.1111111111111113e-05, + "loss": 0.7869, + "step": 55 + }, + { + "epoch": 0.16993835941204363, + "grad_norm": 2.4360105991363525, + "learning_rate": 1.1313131313131314e-05, + "loss": 0.7883, + "step": 56 + }, + { + "epoch": 0.17297297297297298, + "grad_norm": 2.8864669799804688, + "learning_rate": 1.1515151515151517e-05, + "loss": 0.7927, + "step": 57 + }, + { + "epoch": 0.17600758653390233, + "grad_norm": 2.3167998790740967, + "learning_rate": 1.1717171717171718e-05, + "loss": 0.7999, + "step": 58 + }, + { + "epoch": 0.17904220009483168, + "grad_norm": 3.075058698654175, + "learning_rate": 1.191919191919192e-05, + "loss": 0.7888, + "step": 59 + }, + { + "epoch": 0.18207681365576103, + "grad_norm": 2.38899827003479, + "learning_rate": 1.2121212121212122e-05, + "loss": 0.787, + "step": 60 + }, + { + "epoch": 0.18511142721669038, + "grad_norm": 3.121044874191284, + "learning_rate": 1.2323232323232323e-05, + "loss": 0.8053, + "step": 61 + }, + { + "epoch": 0.18814604077761973, + "grad_norm": 2.580725908279419, + "learning_rate": 1.2525252525252527e-05, + "loss": 0.7819, + "step": 62 + }, + { + "epoch": 0.19118065433854908, + "grad_norm": 3.1028575897216797, + "learning_rate": 1.2727272727272728e-05, + "loss": 0.7886, + "step": 63 + }, + { + "epoch": 0.19421526789947843, + "grad_norm": 2.6575424671173096, + "learning_rate": 1.2929292929292931e-05, + "loss": 0.7904, + "step": 64 + }, + { + "epoch": 0.19724988146040778, + "grad_norm": 2.8595755100250244, + "learning_rate": 1.3131313131313132e-05, + "loss": 0.7955, + "step": 65 + }, + { + "epoch": 0.20028449502133713, + "grad_norm": 2.235410451889038, + "learning_rate": 1.3333333333333333e-05, + "loss": 0.796, + "step": 66 + }, + { + "epoch": 0.20331910858226648, + "grad_norm": 2.7857375144958496, + "learning_rate": 1.3535353535353538e-05, + "loss": 0.8008, + "step": 67 + }, + { + "epoch": 0.20635372214319583, + "grad_norm": 2.310331106185913, + "learning_rate": 1.3737373737373739e-05, + "loss": 0.7802, + "step": 68 + }, + { + "epoch": 0.20938833570412518, + "grad_norm": 2.8039300441741943, + "learning_rate": 1.3939393939393942e-05, + "loss": 0.7946, + "step": 69 + }, + { + "epoch": 0.21242294926505453, + "grad_norm": 2.345369815826416, + "learning_rate": 1.4141414141414143e-05, + "loss": 0.7835, + "step": 70 + }, + { + "epoch": 0.21545756282598388, + "grad_norm": 2.8384764194488525, + "learning_rate": 1.4343434343434344e-05, + "loss": 0.7855, + "step": 71 + }, + { + "epoch": 0.21849217638691323, + "grad_norm": 2.371159076690674, + "learning_rate": 1.4545454545454546e-05, + "loss": 0.7689, + "step": 72 + }, + { + "epoch": 0.22152678994784258, + "grad_norm": 2.9578475952148438, + "learning_rate": 1.4747474747474747e-05, + "loss": 0.7918, + "step": 73 + }, + { + "epoch": 0.22456140350877193, + "grad_norm": 2.8900325298309326, + "learning_rate": 1.4949494949494952e-05, + "loss": 0.7746, + "step": 74 + }, + { + "epoch": 0.22759601706970128, + "grad_norm": 2.4469008445739746, + "learning_rate": 1.5151515151515153e-05, + "loss": 0.7919, + "step": 75 + }, + { + "epoch": 0.23063063063063063, + "grad_norm": 2.410154104232788, + "learning_rate": 1.5353535353535354e-05, + "loss": 0.7987, + "step": 76 + }, + { + "epoch": 0.23366524419155998, + "grad_norm": 2.599086284637451, + "learning_rate": 1.555555555555556e-05, + "loss": 0.7714, + "step": 77 + }, + { + "epoch": 0.23669985775248933, + "grad_norm": 2.4329092502593994, + "learning_rate": 1.575757575757576e-05, + "loss": 0.8041, + "step": 78 + }, + { + "epoch": 0.23973447131341868, + "grad_norm": 2.7239413261413574, + "learning_rate": 1.595959595959596e-05, + "loss": 0.7941, + "step": 79 + }, + { + "epoch": 0.24276908487434803, + "grad_norm": 2.8684301376342773, + "learning_rate": 1.616161616161616e-05, + "loss": 0.8076, + "step": 80 + }, + { + "epoch": 0.24580369843527738, + "grad_norm": 3.004608631134033, + "learning_rate": 1.6363636363636366e-05, + "loss": 0.7893, + "step": 81 + }, + { + "epoch": 0.24883831199620673, + "grad_norm": 2.6391353607177734, + "learning_rate": 1.6565656565656567e-05, + "loss": 0.8116, + "step": 82 + }, + { + "epoch": 0.2518729255571361, + "grad_norm": 3.3293023109436035, + "learning_rate": 1.6767676767676768e-05, + "loss": 0.7952, + "step": 83 + }, + { + "epoch": 0.2549075391180654, + "grad_norm": 2.5788087844848633, + "learning_rate": 1.6969696969696972e-05, + "loss": 0.7905, + "step": 84 + }, + { + "epoch": 0.2579421526789948, + "grad_norm": 3.4907588958740234, + "learning_rate": 1.7171717171717173e-05, + "loss": 0.7903, + "step": 85 + }, + { + "epoch": 0.2609767662399241, + "grad_norm": 2.8050403594970703, + "learning_rate": 1.7373737373737375e-05, + "loss": 0.7813, + "step": 86 + }, + { + "epoch": 0.2640113798008535, + "grad_norm": 3.180819034576416, + "learning_rate": 1.7575757575757576e-05, + "loss": 0.7797, + "step": 87 + }, + { + "epoch": 0.26704599336178286, + "grad_norm": 2.614614486694336, + "learning_rate": 1.7777777777777777e-05, + "loss": 0.7903, + "step": 88 + }, + { + "epoch": 0.2700806069227122, + "grad_norm": 3.212218761444092, + "learning_rate": 1.797979797979798e-05, + "loss": 0.7927, + "step": 89 + }, + { + "epoch": 0.27311522048364156, + "grad_norm": 2.6879336833953857, + "learning_rate": 1.8181818181818182e-05, + "loss": 0.8099, + "step": 90 + }, + { + "epoch": 0.2761498340445709, + "grad_norm": 3.2700448036193848, + "learning_rate": 1.8383838383838387e-05, + "loss": 0.794, + "step": 91 + }, + { + "epoch": 0.27918444760550026, + "grad_norm": 3.1249783039093018, + "learning_rate": 1.8585858585858588e-05, + "loss": 0.7807, + "step": 92 + }, + { + "epoch": 0.2822190611664296, + "grad_norm": 2.4789459705352783, + "learning_rate": 1.8787878787878792e-05, + "loss": 0.7829, + "step": 93 + }, + { + "epoch": 0.28525367472735896, + "grad_norm": 2.624887466430664, + "learning_rate": 1.8989898989898993e-05, + "loss": 0.8043, + "step": 94 + }, + { + "epoch": 0.2882882882882883, + "grad_norm": 2.3336539268493652, + "learning_rate": 1.9191919191919194e-05, + "loss": 0.7827, + "step": 95 + }, + { + "epoch": 0.29132290184921766, + "grad_norm": 2.5490119457244873, + "learning_rate": 1.9393939393939395e-05, + "loss": 0.8072, + "step": 96 + }, + { + "epoch": 0.294357515410147, + "grad_norm": 2.5659990310668945, + "learning_rate": 1.9595959595959596e-05, + "loss": 0.7986, + "step": 97 + }, + { + "epoch": 0.29739212897107636, + "grad_norm": 2.3554656505584717, + "learning_rate": 1.97979797979798e-05, + "loss": 0.8059, + "step": 98 + }, + { + "epoch": 0.3004267425320057, + "grad_norm": 2.4364328384399414, + "learning_rate": 2e-05, + "loss": 0.7874, + "step": 99 + }, + { + "epoch": 0.30346135609293506, + "grad_norm": 2.3030965328216553, + "learning_rate": 1.9999937418875125e-05, + "loss": 0.7854, + "step": 100 + }, + { + "epoch": 0.30346135609293506, + "eval_loss": NaN, + "eval_runtime": 204.7053, + "eval_samples_per_second": 20.312, + "eval_steps_per_second": 5.08, + "step": 100 + }, + { + "epoch": 0.3064959696538644, + "grad_norm": 2.318124294281006, + "learning_rate": 1.9999749676283775e-05, + "loss": 0.7964, + "step": 101 + }, + { + "epoch": 0.30953058321479376, + "grad_norm": 2.2906177043914795, + "learning_rate": 1.9999436774575783e-05, + "loss": 0.8049, + "step": 102 + }, + { + "epoch": 0.3125651967757231, + "grad_norm": 2.2947778701782227, + "learning_rate": 1.999899871766749e-05, + "loss": 0.7881, + "step": 103 + }, + { + "epoch": 0.31559981033665246, + "grad_norm": 2.548788547515869, + "learning_rate": 1.999843551104172e-05, + "loss": 0.803, + "step": 104 + }, + { + "epoch": 0.3186344238975818, + "grad_norm": 2.3735954761505127, + "learning_rate": 1.9997747161747696e-05, + "loss": 0.8044, + "step": 105 + }, + { + "epoch": 0.32166903745851116, + "grad_norm": 2.620364189147949, + "learning_rate": 1.9996933678400948e-05, + "loss": 0.7806, + "step": 106 + }, + { + "epoch": 0.3247036510194405, + "grad_norm": 2.3897509574890137, + "learning_rate": 1.999599507118322e-05, + "loss": 0.7862, + "step": 107 + }, + { + "epoch": 0.32773826458036986, + "grad_norm": 2.6068966388702393, + "learning_rate": 1.9994931351842327e-05, + "loss": 0.8051, + "step": 108 + }, + { + "epoch": 0.3307728781412992, + "grad_norm": 2.560683488845825, + "learning_rate": 1.999374253369202e-05, + "loss": 0.7956, + "step": 109 + }, + { + "epoch": 0.33380749170222856, + "grad_norm": 2.1629045009613037, + "learning_rate": 1.999242863161182e-05, + "loss": 0.8074, + "step": 110 + }, + { + "epoch": 0.3368421052631579, + "grad_norm": 2.653731346130371, + "learning_rate": 1.999098966204682e-05, + "loss": 0.8071, + "step": 111 + }, + { + "epoch": 0.33987671882408726, + "grad_norm": 2.484330177307129, + "learning_rate": 1.9989425643007475e-05, + "loss": 0.8135, + "step": 112 + }, + { + "epoch": 0.3429113323850166, + "grad_norm": 2.3897571563720703, + "learning_rate": 1.9987736594069417e-05, + "loss": 0.8066, + "step": 113 + }, + { + "epoch": 0.34594594594594597, + "grad_norm": 2.653904676437378, + "learning_rate": 1.998592253637315e-05, + "loss": 0.7908, + "step": 114 + }, + { + "epoch": 0.3489805595068753, + "grad_norm": 2.2755186557769775, + "learning_rate": 1.9983983492623832e-05, + "loss": 0.7824, + "step": 115 + }, + { + "epoch": 0.35201517306780467, + "grad_norm": 2.260126829147339, + "learning_rate": 1.9981919487090973e-05, + "loss": 0.7916, + "step": 116 + }, + { + "epoch": 0.355049786628734, + "grad_norm": 2.6538352966308594, + "learning_rate": 1.9979730545608128e-05, + "loss": 0.7927, + "step": 117 + }, + { + "epoch": 0.35808440018966337, + "grad_norm": 2.2571558952331543, + "learning_rate": 1.9977416695572577e-05, + "loss": 0.7826, + "step": 118 + }, + { + "epoch": 0.3611190137505927, + "grad_norm": 2.5154271125793457, + "learning_rate": 1.9974977965945e-05, + "loss": 0.807, + "step": 119 + }, + { + "epoch": 0.36415362731152207, + "grad_norm": 2.4908971786499023, + "learning_rate": 1.9972414387249074e-05, + "loss": 0.7831, + "step": 120 + }, + { + "epoch": 0.3671882408724514, + "grad_norm": 2.3925859928131104, + "learning_rate": 1.996972599157113e-05, + "loss": 0.7844, + "step": 121 + }, + { + "epoch": 0.37022285443338077, + "grad_norm": 2.3527793884277344, + "learning_rate": 1.9966912812559733e-05, + "loss": 0.7921, + "step": 122 + }, + { + "epoch": 0.3732574679943101, + "grad_norm": 2.2694365978240967, + "learning_rate": 1.9963974885425267e-05, + "loss": 0.7816, + "step": 123 + }, + { + "epoch": 0.37629208155523947, + "grad_norm": 2.436676025390625, + "learning_rate": 1.9960912246939486e-05, + "loss": 0.7782, + "step": 124 + }, + { + "epoch": 0.3793266951161688, + "grad_norm": 2.3587653636932373, + "learning_rate": 1.9957724935435065e-05, + "loss": 0.8024, + "step": 125 + }, + { + "epoch": 0.38236130867709817, + "grad_norm": 2.3145172595977783, + "learning_rate": 1.9954412990805107e-05, + "loss": 0.8115, + "step": 126 + }, + { + "epoch": 0.3853959222380275, + "grad_norm": 2.271946430206299, + "learning_rate": 1.995097645450266e-05, + "loss": 0.7975, + "step": 127 + }, + { + "epoch": 0.38843053579895687, + "grad_norm": 2.2533860206604004, + "learning_rate": 1.994741536954019e-05, + "loss": 0.8187, + "step": 128 + }, + { + "epoch": 0.3914651493598862, + "grad_norm": 2.5504581928253174, + "learning_rate": 1.994372978048903e-05, + "loss": 0.7913, + "step": 129 + }, + { + "epoch": 0.39449976292081557, + "grad_norm": 2.3467888832092285, + "learning_rate": 1.993991973347884e-05, + "loss": 0.7955, + "step": 130 + }, + { + "epoch": 0.3975343764817449, + "grad_norm": 2.356933832168579, + "learning_rate": 1.9935985276197033e-05, + "loss": 0.808, + "step": 131 + }, + { + "epoch": 0.40056899004267427, + "grad_norm": 2.608546257019043, + "learning_rate": 1.9931926457888155e-05, + "loss": 0.785, + "step": 132 + }, + { + "epoch": 0.4036036036036036, + "grad_norm": 2.2730495929718018, + "learning_rate": 1.9927743329353295e-05, + "loss": 0.79, + "step": 133 + }, + { + "epoch": 0.40663821716453297, + "grad_norm": 2.2720224857330322, + "learning_rate": 1.992343594294943e-05, + "loss": 0.8084, + "step": 134 + }, + { + "epoch": 0.4096728307254623, + "grad_norm": 2.2325122356414795, + "learning_rate": 1.9919004352588768e-05, + "loss": 0.8008, + "step": 135 + }, + { + "epoch": 0.41270744428639167, + "grad_norm": 2.4259414672851562, + "learning_rate": 1.9914448613738107e-05, + "loss": 0.7827, + "step": 136 + }, + { + "epoch": 0.415742057847321, + "grad_norm": 2.2663819789886475, + "learning_rate": 1.9909768783418086e-05, + "loss": 0.8059, + "step": 137 + }, + { + "epoch": 0.41877667140825037, + "grad_norm": 2.910830020904541, + "learning_rate": 1.990496492020252e-05, + "loss": 0.8159, + "step": 138 + }, + { + "epoch": 0.4218112849691797, + "grad_norm": 2.4485902786254883, + "learning_rate": 1.9900037084217637e-05, + "loss": 0.7921, + "step": 139 + }, + { + "epoch": 0.42484589853010907, + "grad_norm": 2.3199424743652344, + "learning_rate": 1.989498533714135e-05, + "loss": 0.8006, + "step": 140 + }, + { + "epoch": 0.4278805120910384, + "grad_norm": 2.529900550842285, + "learning_rate": 1.9889809742202454e-05, + "loss": 0.7805, + "step": 141 + }, + { + "epoch": 0.43091512565196777, + "grad_norm": 2.5135438442230225, + "learning_rate": 1.988451036417986e-05, + "loss": 0.8088, + "step": 142 + }, + { + "epoch": 0.4339497392128971, + "grad_norm": 2.359200954437256, + "learning_rate": 1.9879087269401782e-05, + "loss": 0.7963, + "step": 143 + }, + { + "epoch": 0.43698435277382647, + "grad_norm": 2.307164192199707, + "learning_rate": 1.9873540525744888e-05, + "loss": 0.79, + "step": 144 + }, + { + "epoch": 0.4400189663347558, + "grad_norm": 2.341834545135498, + "learning_rate": 1.986787020263347e-05, + "loss": 0.7955, + "step": 145 + }, + { + "epoch": 0.44305357989568517, + "grad_norm": 2.409449338912964, + "learning_rate": 1.986207637103857e-05, + "loss": 0.7761, + "step": 146 + }, + { + "epoch": 0.4460881934566145, + "grad_norm": 2.3119232654571533, + "learning_rate": 1.9856159103477085e-05, + "loss": 0.7992, + "step": 147 + }, + { + "epoch": 0.44912280701754387, + "grad_norm": 2.3376715183258057, + "learning_rate": 1.9850118474010873e-05, + "loss": 0.784, + "step": 148 + }, + { + "epoch": 0.4521574205784732, + "grad_norm": 2.387392997741699, + "learning_rate": 1.98439545582458e-05, + "loss": 0.7806, + "step": 149 + }, + { + "epoch": 0.45519203413940257, + "grad_norm": 2.2223150730133057, + "learning_rate": 1.983766743333084e-05, + "loss": 0.7914, + "step": 150 + }, + { + "epoch": 0.45519203413940257, + "eval_loss": NaN, + "eval_runtime": 205.0592, + "eval_samples_per_second": 20.277, + "eval_steps_per_second": 5.072, + "step": 150 + }, + { + "epoch": 0.4582266477003319, + "grad_norm": 2.727046012878418, + "learning_rate": 1.9831257177957045e-05, + "loss": 0.793, + "step": 151 + }, + { + "epoch": 0.46126126126126127, + "grad_norm": 2.266788959503174, + "learning_rate": 1.9824723872356623e-05, + "loss": 0.8127, + "step": 152 + }, + { + "epoch": 0.4642958748221906, + "grad_norm": 2.530904531478882, + "learning_rate": 1.9818067598301894e-05, + "loss": 0.7973, + "step": 153 + }, + { + "epoch": 0.46733048838311997, + "grad_norm": 2.3305609226226807, + "learning_rate": 1.981128843910428e-05, + "loss": 0.7961, + "step": 154 + }, + { + "epoch": 0.4703651019440493, + "grad_norm": 2.1240079402923584, + "learning_rate": 1.9804386479613268e-05, + "loss": 0.8093, + "step": 155 + }, + { + "epoch": 0.47339971550497867, + "grad_norm": 2.625185489654541, + "learning_rate": 1.9797361806215335e-05, + "loss": 0.7988, + "step": 156 + }, + { + "epoch": 0.476434329065908, + "grad_norm": 2.4407734870910645, + "learning_rate": 1.9790214506832868e-05, + "loss": 0.8166, + "step": 157 + }, + { + "epoch": 0.47946894262683737, + "grad_norm": 2.4013476371765137, + "learning_rate": 1.9782944670923075e-05, + "loss": 0.7935, + "step": 158 + }, + { + "epoch": 0.4825035561877667, + "grad_norm": 2.357010841369629, + "learning_rate": 1.9775552389476865e-05, + "loss": 0.7818, + "step": 159 + }, + { + "epoch": 0.48553816974869607, + "grad_norm": 2.1862401962280273, + "learning_rate": 1.9768037755017687e-05, + "loss": 0.7879, + "step": 160 + }, + { + "epoch": 0.4885727833096254, + "grad_norm": 2.3130927085876465, + "learning_rate": 1.97604008616004e-05, + "loss": 0.7914, + "step": 161 + }, + { + "epoch": 0.49160739687055477, + "grad_norm": 2.0661509037017822, + "learning_rate": 1.9752641804810083e-05, + "loss": 0.8023, + "step": 162 + }, + { + "epoch": 0.4946420104314841, + "grad_norm": 2.2117955684661865, + "learning_rate": 1.9744760681760832e-05, + "loss": 0.7972, + "step": 163 + }, + { + "epoch": 0.49767662399241347, + "grad_norm": 2.603163242340088, + "learning_rate": 1.973675759109456e-05, + "loss": 0.7913, + "step": 164 + }, + { + "epoch": 0.5007112375533428, + "grad_norm": 2.2511062622070312, + "learning_rate": 1.9728632632979746e-05, + "loss": 0.7914, + "step": 165 + }, + { + "epoch": 0.5037458511142722, + "grad_norm": 2.375213861465454, + "learning_rate": 1.9720385909110197e-05, + "loss": 0.7928, + "step": 166 + }, + { + "epoch": 0.5067804646752015, + "grad_norm": 2.5348660945892334, + "learning_rate": 1.9712017522703764e-05, + "loss": 0.7894, + "step": 167 + }, + { + "epoch": 0.5098150782361308, + "grad_norm": 2.2094035148620605, + "learning_rate": 1.9703527578501052e-05, + "loss": 0.7813, + "step": 168 + }, + { + "epoch": 0.5128496917970602, + "grad_norm": 2.2283380031585693, + "learning_rate": 1.9694916182764113e-05, + "loss": 0.7877, + "step": 169 + }, + { + "epoch": 0.5158843053579896, + "grad_norm": 2.189119338989258, + "learning_rate": 1.9686183443275118e-05, + "loss": 0.7989, + "step": 170 + }, + { + "epoch": 0.518918918918919, + "grad_norm": 2.2636640071868896, + "learning_rate": 1.967732946933499e-05, + "loss": 0.8059, + "step": 171 + }, + { + "epoch": 0.5219535324798482, + "grad_norm": 2.243251085281372, + "learning_rate": 1.9668354371762066e-05, + "loss": 0.7904, + "step": 172 + }, + { + "epoch": 0.5249881460407776, + "grad_norm": 2.215536117553711, + "learning_rate": 1.9659258262890683e-05, + "loss": 0.7912, + "step": 173 + }, + { + "epoch": 0.528022759601707, + "grad_norm": 2.2998199462890625, + "learning_rate": 1.9650041256569792e-05, + "loss": 0.797, + "step": 174 + }, + { + "epoch": 0.5310573731626363, + "grad_norm": 2.3586950302124023, + "learning_rate": 1.9640703468161508e-05, + "loss": 0.7907, + "step": 175 + }, + { + "epoch": 0.5340919867235657, + "grad_norm": 2.257404088973999, + "learning_rate": 1.96312450145397e-05, + "loss": 0.7977, + "step": 176 + }, + { + "epoch": 0.537126600284495, + "grad_norm": 2.4808526039123535, + "learning_rate": 1.9621666014088495e-05, + "loss": 0.7929, + "step": 177 + }, + { + "epoch": 0.5401612138454244, + "grad_norm": 2.208704710006714, + "learning_rate": 1.9611966586700825e-05, + "loss": 0.7975, + "step": 178 + }, + { + "epoch": 0.5431958274063537, + "grad_norm": 2.4615161418914795, + "learning_rate": 1.9602146853776894e-05, + "loss": 0.7991, + "step": 179 + }, + { + "epoch": 0.5462304409672831, + "grad_norm": 2.4766757488250732, + "learning_rate": 1.9592206938222703e-05, + "loss": 0.7911, + "step": 180 + }, + { + "epoch": 0.5492650545282124, + "grad_norm": 2.2009432315826416, + "learning_rate": 1.9582146964448457e-05, + "loss": 0.788, + "step": 181 + }, + { + "epoch": 0.5522996680891418, + "grad_norm": 2.258129358291626, + "learning_rate": 1.9571967058367067e-05, + "loss": 0.7893, + "step": 182 + }, + { + "epoch": 0.5553342816500711, + "grad_norm": 2.1914985179901123, + "learning_rate": 1.956166734739251e-05, + "loss": 0.8057, + "step": 183 + }, + { + "epoch": 0.5583688952110005, + "grad_norm": 2.2815279960632324, + "learning_rate": 1.9551247960438298e-05, + "loss": 0.7823, + "step": 184 + }, + { + "epoch": 0.5614035087719298, + "grad_norm": 2.2393579483032227, + "learning_rate": 1.954070902791582e-05, + "loss": 0.7899, + "step": 185 + }, + { + "epoch": 0.5644381223328592, + "grad_norm": 2.56640625, + "learning_rate": 1.953005068173272e-05, + "loss": 0.7731, + "step": 186 + }, + { + "epoch": 0.5674727358937885, + "grad_norm": 2.3896234035491943, + "learning_rate": 1.9519273055291266e-05, + "loss": 0.7936, + "step": 187 + }, + { + "epoch": 0.5705073494547179, + "grad_norm": 2.21494722366333, + "learning_rate": 1.9508376283486653e-05, + "loss": 0.8121, + "step": 188 + }, + { + "epoch": 0.5735419630156472, + "grad_norm": 2.400538444519043, + "learning_rate": 1.949736050270532e-05, + "loss": 0.7938, + "step": 189 + }, + { + "epoch": 0.5765765765765766, + "grad_norm": 2.1337621212005615, + "learning_rate": 1.9486225850823265e-05, + "loss": 0.8049, + "step": 190 + }, + { + "epoch": 0.579611190137506, + "grad_norm": 2.314168930053711, + "learning_rate": 1.9474972467204298e-05, + "loss": 0.8109, + "step": 191 + }, + { + "epoch": 0.5826458036984353, + "grad_norm": 2.2364132404327393, + "learning_rate": 1.9463600492698297e-05, + "loss": 0.78, + "step": 192 + }, + { + "epoch": 0.5856804172593646, + "grad_norm": 2.204206705093384, + "learning_rate": 1.945211006963945e-05, + "loss": 0.8187, + "step": 193 + }, + { + "epoch": 0.588715030820294, + "grad_norm": 2.1865625381469727, + "learning_rate": 1.9440501341844484e-05, + "loss": 0.7844, + "step": 194 + }, + { + "epoch": 0.5917496443812233, + "grad_norm": 2.143092155456543, + "learning_rate": 1.9428774454610845e-05, + "loss": 0.7967, + "step": 195 + }, + { + "epoch": 0.5947842579421527, + "grad_norm": 2.3440985679626465, + "learning_rate": 1.9416929554714887e-05, + "loss": 0.8096, + "step": 196 + }, + { + "epoch": 0.597818871503082, + "grad_norm": 2.43977689743042, + "learning_rate": 1.9404966790410047e-05, + "loss": 0.795, + "step": 197 + }, + { + "epoch": 0.6008534850640114, + "grad_norm": 2.2740390300750732, + "learning_rate": 1.9392886311424975e-05, + "loss": 0.7937, + "step": 198 + }, + { + "epoch": 0.6038880986249408, + "grad_norm": 2.1775126457214355, + "learning_rate": 1.938068826896166e-05, + "loss": 0.8019, + "step": 199 + }, + { + "epoch": 0.6069227121858701, + "grad_norm": 2.1858487129211426, + "learning_rate": 1.9368372815693547e-05, + "loss": 0.7962, + "step": 200 + }, + { + "epoch": 0.6069227121858701, + "eval_loss": NaN, + "eval_runtime": 204.6421, + "eval_samples_per_second": 20.318, + "eval_steps_per_second": 5.082, + "step": 200 + }, + { + "epoch": 0.6099573257467994, + "grad_norm": 2.2958755493164062, + "learning_rate": 1.9355940105763622e-05, + "loss": 0.8003, + "step": 201 + }, + { + "epoch": 0.6129919393077288, + "grad_norm": 2.1330931186676025, + "learning_rate": 1.934339029478248e-05, + "loss": 0.7803, + "step": 202 + }, + { + "epoch": 0.6160265528686582, + "grad_norm": 2.3881402015686035, + "learning_rate": 1.9330723539826373e-05, + "loss": 0.8044, + "step": 203 + }, + { + "epoch": 0.6190611664295875, + "grad_norm": 2.2404513359069824, + "learning_rate": 1.9317939999435262e-05, + "loss": 0.8097, + "step": 204 + }, + { + "epoch": 0.6220957799905168, + "grad_norm": 2.194645404815674, + "learning_rate": 1.930503983361081e-05, + "loss": 0.8178, + "step": 205 + }, + { + "epoch": 0.6251303935514462, + "grad_norm": 2.509723424911499, + "learning_rate": 1.92920232038144e-05, + "loss": 0.7936, + "step": 206 + }, + { + "epoch": 0.6281650071123756, + "grad_norm": 2.2457869052886963, + "learning_rate": 1.9278890272965097e-05, + "loss": 0.8162, + "step": 207 + }, + { + "epoch": 0.6311996206733049, + "grad_norm": 2.2989683151245117, + "learning_rate": 1.9265641205437612e-05, + "loss": 0.8012, + "step": 208 + }, + { + "epoch": 0.6342342342342342, + "grad_norm": 2.3188092708587646, + "learning_rate": 1.925227616706026e-05, + "loss": 0.8094, + "step": 209 + }, + { + "epoch": 0.6372688477951636, + "grad_norm": 2.2978076934814453, + "learning_rate": 1.9238795325112867e-05, + "loss": 0.7811, + "step": 210 + }, + { + "epoch": 0.640303461356093, + "grad_norm": 2.331542730331421, + "learning_rate": 1.9225198848324687e-05, + "loss": 0.7946, + "step": 211 + }, + { + "epoch": 0.6433380749170223, + "grad_norm": 2.1989738941192627, + "learning_rate": 1.921148690687228e-05, + "loss": 0.7983, + "step": 212 + }, + { + "epoch": 0.6463726884779516, + "grad_norm": 2.4261419773101807, + "learning_rate": 1.9197659672377388e-05, + "loss": 0.8134, + "step": 213 + }, + { + "epoch": 0.649407302038881, + "grad_norm": 2.3790082931518555, + "learning_rate": 1.918371731790479e-05, + "loss": 0.79, + "step": 214 + }, + { + "epoch": 0.6524419155998104, + "grad_norm": 2.2966949939727783, + "learning_rate": 1.9169660017960135e-05, + "loss": 0.8029, + "step": 215 + }, + { + "epoch": 0.6554765291607397, + "grad_norm": 2.5911426544189453, + "learning_rate": 1.915548794848775e-05, + "loss": 0.8118, + "step": 216 + }, + { + "epoch": 0.658511142721669, + "grad_norm": 2.3282856941223145, + "learning_rate": 1.9141201286868435e-05, + "loss": 0.8092, + "step": 217 + }, + { + "epoch": 0.6615457562825984, + "grad_norm": 2.415398359298706, + "learning_rate": 1.9126800211917277e-05, + "loss": 0.8156, + "step": 218 + }, + { + "epoch": 0.6645803698435278, + "grad_norm": 2.2823410034179688, + "learning_rate": 1.911228490388136e-05, + "loss": 0.8004, + "step": 219 + }, + { + "epoch": 0.6676149834044571, + "grad_norm": 2.2104527950286865, + "learning_rate": 1.9097655544437544e-05, + "loss": 0.8023, + "step": 220 + }, + { + "epoch": 0.6706495969653864, + "grad_norm": 2.3354063034057617, + "learning_rate": 1.908291231669019e-05, + "loss": 0.8117, + "step": 221 + }, + { + "epoch": 0.6736842105263158, + "grad_norm": 2.230656147003174, + "learning_rate": 1.906805540516885e-05, + "loss": 0.7797, + "step": 222 + }, + { + "epoch": 0.6767188240872452, + "grad_norm": 2.464111328125, + "learning_rate": 1.905308499582597e-05, + "loss": 0.7929, + "step": 223 + }, + { + "epoch": 0.6797534376481745, + "grad_norm": 2.188788652420044, + "learning_rate": 1.903800127603456e-05, + "loss": 0.7971, + "step": 224 + }, + { + "epoch": 0.6827880512091038, + "grad_norm": 2.202427387237549, + "learning_rate": 1.9022804434585854e-05, + "loss": 0.8026, + "step": 225 + }, + { + "epoch": 0.6858226647700332, + "grad_norm": 2.2621190547943115, + "learning_rate": 1.9007494661686937e-05, + "loss": 0.8112, + "step": 226 + }, + { + "epoch": 0.6888572783309626, + "grad_norm": 2.333603620529175, + "learning_rate": 1.8992072148958368e-05, + "loss": 0.7937, + "step": 227 + }, + { + "epoch": 0.6918918918918919, + "grad_norm": 2.626451253890991, + "learning_rate": 1.8976537089431793e-05, + "loss": 0.8005, + "step": 228 + }, + { + "epoch": 0.6949265054528212, + "grad_norm": 2.4107227325439453, + "learning_rate": 1.8960889677547506e-05, + "loss": 0.7813, + "step": 229 + }, + { + "epoch": 0.6979611190137506, + "grad_norm": 2.7147607803344727, + "learning_rate": 1.8945130109152035e-05, + "loss": 0.8036, + "step": 230 + }, + { + "epoch": 0.70099573257468, + "grad_norm": 2.281703233718872, + "learning_rate": 1.8929258581495688e-05, + "loss": 0.7946, + "step": 231 + }, + { + "epoch": 0.7040303461356093, + "grad_norm": 3.016942262649536, + "learning_rate": 1.891327529323007e-05, + "loss": 0.7786, + "step": 232 + }, + { + "epoch": 0.7070649596965386, + "grad_norm": 2.5729317665100098, + "learning_rate": 1.8897180444405615e-05, + "loss": 0.8141, + "step": 233 + }, + { + "epoch": 0.710099573257468, + "grad_norm": 2.75722336769104, + "learning_rate": 1.888097423646907e-05, + "loss": 0.8079, + "step": 234 + }, + { + "epoch": 0.7131341868183974, + "grad_norm": 2.843980073928833, + "learning_rate": 1.8864656872260985e-05, + "loss": 0.795, + "step": 235 + }, + { + "epoch": 0.7161688003793267, + "grad_norm": 3.3999879360198975, + "learning_rate": 1.884822855601316e-05, + "loss": 0.8086, + "step": 236 + }, + { + "epoch": 0.719203413940256, + "grad_norm": 3.1997487545013428, + "learning_rate": 1.8831689493346095e-05, + "loss": 0.7919, + "step": 237 + }, + { + "epoch": 0.7222380275011854, + "grad_norm": 2.4851367473602295, + "learning_rate": 1.881503989126642e-05, + "loss": 0.7914, + "step": 238 + }, + { + "epoch": 0.7252726410621148, + "grad_norm": 2.738428831100464, + "learning_rate": 1.8798279958164295e-05, + "loss": 0.7929, + "step": 239 + }, + { + "epoch": 0.7283072546230441, + "grad_norm": 2.1467783451080322, + "learning_rate": 1.8781409903810823e-05, + "loss": 0.815, + "step": 240 + }, + { + "epoch": 0.7313418681839734, + "grad_norm": 16.773521423339844, + "learning_rate": 1.8764429939355394e-05, + "loss": 0.8014, + "step": 241 + }, + { + "epoch": 0.7343764817449028, + "grad_norm": 3.171633720397949, + "learning_rate": 1.874734027732306e-05, + "loss": 0.805, + "step": 242 + }, + { + "epoch": 0.7374110953058322, + "grad_norm": 2.249514102935791, + "learning_rate": 1.8730141131611882e-05, + "loss": 0.8109, + "step": 243 + }, + { + "epoch": 0.7404457088667615, + "grad_norm": 2.816488742828369, + "learning_rate": 1.8712832717490238e-05, + "loss": 0.8044, + "step": 244 + }, + { + "epoch": 0.7434803224276908, + "grad_norm": 2.558295488357544, + "learning_rate": 1.8695415251594123e-05, + "loss": 0.8269, + "step": 245 + }, + { + "epoch": 0.7465149359886202, + "grad_norm": 2.5777034759521484, + "learning_rate": 1.8677888951924473e-05, + "loss": 0.7971, + "step": 246 + }, + { + "epoch": 0.7495495495495496, + "grad_norm": 2.394287586212158, + "learning_rate": 1.866025403784439e-05, + "loss": 0.8092, + "step": 247 + }, + { + "epoch": 0.7525841631104789, + "grad_norm": 2.2798614501953125, + "learning_rate": 1.864251073007642e-05, + "loss": 0.7964, + "step": 248 + }, + { + "epoch": 0.7556187766714082, + "grad_norm": 2.3587262630462646, + "learning_rate": 1.8624659250699807e-05, + "loss": 0.7928, + "step": 249 + }, + { + "epoch": 0.7586533902323376, + "grad_norm": 2.189763307571411, + "learning_rate": 1.8606699823147675e-05, + "loss": 0.7812, + "step": 250 + }, + { + "epoch": 0.7586533902323376, + "eval_loss": NaN, + "eval_runtime": 204.0596, + "eval_samples_per_second": 20.376, + "eval_steps_per_second": 5.097, + "step": 250 + }, + { + "epoch": 0.761688003793267, + "grad_norm": 2.433803081512451, + "learning_rate": 1.8588632672204264e-05, + "loss": 0.8111, + "step": 251 + }, + { + "epoch": 0.7647226173541963, + "grad_norm": 2.28267765045166, + "learning_rate": 1.8570458024002094e-05, + "loss": 0.8001, + "step": 252 + }, + { + "epoch": 0.7677572309151256, + "grad_norm": 2.3339545726776123, + "learning_rate": 1.8552176106019156e-05, + "loss": 0.8158, + "step": 253 + }, + { + "epoch": 0.770791844476055, + "grad_norm": 2.284759998321533, + "learning_rate": 1.8533787147076046e-05, + "loss": 0.7852, + "step": 254 + }, + { + "epoch": 0.7738264580369844, + "grad_norm": 2.35969614982605, + "learning_rate": 1.8515291377333114e-05, + "loss": 0.7909, + "step": 255 + }, + { + "epoch": 0.7768610715979137, + "grad_norm": 2.4525341987609863, + "learning_rate": 1.8496689028287572e-05, + "loss": 0.817, + "step": 256 + }, + { + "epoch": 0.779895685158843, + "grad_norm": 2.7126755714416504, + "learning_rate": 1.847798033277061e-05, + "loss": 0.797, + "step": 257 + }, + { + "epoch": 0.7829302987197724, + "grad_norm": 2.2292561531066895, + "learning_rate": 1.8459165524944463e-05, + "loss": 0.8044, + "step": 258 + }, + { + "epoch": 0.7859649122807018, + "grad_norm": 2.447347640991211, + "learning_rate": 1.8440244840299507e-05, + "loss": 0.7979, + "step": 259 + }, + { + "epoch": 0.7889995258416311, + "grad_norm": 2.2336087226867676, + "learning_rate": 1.842121851565128e-05, + "loss": 0.8036, + "step": 260 + }, + { + "epoch": 0.7920341394025604, + "grad_norm": 2.3720502853393555, + "learning_rate": 1.8402086789137547e-05, + "loss": 0.7979, + "step": 261 + }, + { + "epoch": 0.7950687529634898, + "grad_norm": 2.1780807971954346, + "learning_rate": 1.8382849900215297e-05, + "loss": 0.7876, + "step": 262 + }, + { + "epoch": 0.7981033665244192, + "grad_norm": 2.3325858116149902, + "learning_rate": 1.8363508089657763e-05, + "loss": 0.7997, + "step": 263 + }, + { + "epoch": 0.8011379800853485, + "grad_norm": 2.3341164588928223, + "learning_rate": 1.8344061599551397e-05, + "loss": 0.7844, + "step": 264 + }, + { + "epoch": 0.8041725936462779, + "grad_norm": 2.4678280353546143, + "learning_rate": 1.8324510673292844e-05, + "loss": 0.7946, + "step": 265 + }, + { + "epoch": 0.8072072072072072, + "grad_norm": 2.424893856048584, + "learning_rate": 1.8304855555585893e-05, + "loss": 0.7916, + "step": 266 + }, + { + "epoch": 0.8102418207681366, + "grad_norm": 2.5433976650238037, + "learning_rate": 1.8285096492438424e-05, + "loss": 0.7983, + "step": 267 + }, + { + "epoch": 0.8132764343290659, + "grad_norm": 2.499178647994995, + "learning_rate": 1.826523373115931e-05, + "loss": 0.7944, + "step": 268 + }, + { + "epoch": 0.8163110478899953, + "grad_norm": 2.2453994750976562, + "learning_rate": 1.8245267520355348e-05, + "loss": 0.8148, + "step": 269 + }, + { + "epoch": 0.8193456614509246, + "grad_norm": 2.350146770477295, + "learning_rate": 1.8225198109928116e-05, + "loss": 0.7986, + "step": 270 + }, + { + "epoch": 0.822380275011854, + "grad_norm": 2.2048559188842773, + "learning_rate": 1.8205025751070878e-05, + "loss": 0.8093, + "step": 271 + }, + { + "epoch": 0.8254148885727833, + "grad_norm": 2.2730185985565186, + "learning_rate": 1.8184750696265408e-05, + "loss": 0.7787, + "step": 272 + }, + { + "epoch": 0.8284495021337127, + "grad_norm": 2.423301935195923, + "learning_rate": 1.8164373199278858e-05, + "loss": 0.823, + "step": 273 + }, + { + "epoch": 0.831484115694642, + "grad_norm": 2.309649705886841, + "learning_rate": 1.8143893515160565e-05, + "loss": 0.7901, + "step": 274 + }, + { + "epoch": 0.8345187292555714, + "grad_norm": 2.249284267425537, + "learning_rate": 1.812331190023886e-05, + "loss": 0.8095, + "step": 275 + }, + { + "epoch": 0.8375533428165007, + "grad_norm": 2.2063703536987305, + "learning_rate": 1.8102628612117868e-05, + "loss": 0.8008, + "step": 276 + }, + { + "epoch": 0.8405879563774301, + "grad_norm": 2.2518839836120605, + "learning_rate": 1.8081843909674277e-05, + "loss": 0.8051, + "step": 277 + }, + { + "epoch": 0.8436225699383594, + "grad_norm": 2.226356267929077, + "learning_rate": 1.8060958053054095e-05, + "loss": 0.8036, + "step": 278 + }, + { + "epoch": 0.8466571834992888, + "grad_norm": 2.186485767364502, + "learning_rate": 1.8039971303669407e-05, + "loss": 0.8025, + "step": 279 + }, + { + "epoch": 0.8496917970602181, + "grad_norm": 2.4235646724700928, + "learning_rate": 1.8018883924195085e-05, + "loss": 0.7799, + "step": 280 + }, + { + "epoch": 0.8527264106211475, + "grad_norm": 2.2470943927764893, + "learning_rate": 1.799769617856552e-05, + "loss": 0.8025, + "step": 281 + }, + { + "epoch": 0.8557610241820768, + "grad_norm": 2.2514889240264893, + "learning_rate": 1.79764083319713e-05, + "loss": 0.7978, + "step": 282 + }, + { + "epoch": 0.8587956377430062, + "grad_norm": 2.220952033996582, + "learning_rate": 1.79550206508559e-05, + "loss": 0.7978, + "step": 283 + }, + { + "epoch": 0.8618302513039355, + "grad_norm": 2.4699270725250244, + "learning_rate": 1.7933533402912354e-05, + "loss": 0.7835, + "step": 284 + }, + { + "epoch": 0.8648648648648649, + "grad_norm": 2.3011207580566406, + "learning_rate": 1.7911946857079886e-05, + "loss": 0.8008, + "step": 285 + }, + { + "epoch": 0.8678994784257942, + "grad_norm": 2.2239327430725098, + "learning_rate": 1.7890261283540563e-05, + "loss": 0.7904, + "step": 286 + }, + { + "epoch": 0.8709340919867236, + "grad_norm": 2.077845573425293, + "learning_rate": 1.78684769537159e-05, + "loss": 0.7962, + "step": 287 + }, + { + "epoch": 0.8739687055476529, + "grad_norm": 2.2492687702178955, + "learning_rate": 1.7846594140263475e-05, + "loss": 0.8076, + "step": 288 + }, + { + "epoch": 0.8770033191085823, + "grad_norm": 2.1000773906707764, + "learning_rate": 1.78246131170735e-05, + "loss": 0.8107, + "step": 289 + }, + { + "epoch": 0.8800379326695116, + "grad_norm": 2.2440242767333984, + "learning_rate": 1.7802534159265407e-05, + "loss": 0.784, + "step": 290 + }, + { + "epoch": 0.883072546230441, + "grad_norm": 2.1370065212249756, + "learning_rate": 1.7780357543184396e-05, + "loss": 0.7926, + "step": 291 + }, + { + "epoch": 0.8861071597913703, + "grad_norm": 2.2182703018188477, + "learning_rate": 1.775808354639799e-05, + "loss": 0.8079, + "step": 292 + }, + { + "epoch": 0.8891417733522997, + "grad_norm": 2.236370325088501, + "learning_rate": 1.773571244769254e-05, + "loss": 0.7851, + "step": 293 + }, + { + "epoch": 0.892176386913229, + "grad_norm": 2.216042995452881, + "learning_rate": 1.771324452706975e-05, + "loss": 0.798, + "step": 294 + }, + { + "epoch": 0.8952110004741584, + "grad_norm": 2.2804715633392334, + "learning_rate": 1.769068006574317e-05, + "loss": 0.7916, + "step": 295 + }, + { + "epoch": 0.8982456140350877, + "grad_norm": 2.188271999359131, + "learning_rate": 1.7668019346134674e-05, + "loss": 0.7993, + "step": 296 + }, + { + "epoch": 0.9012802275960171, + "grad_norm": 2.372596025466919, + "learning_rate": 1.7645262651870926e-05, + "loss": 0.816, + "step": 297 + }, + { + "epoch": 0.9043148411569464, + "grad_norm": 2.174302339553833, + "learning_rate": 1.7622410267779834e-05, + "loss": 0.8247, + "step": 298 + }, + { + "epoch": 0.9073494547178758, + "grad_norm": 2.0944302082061768, + "learning_rate": 1.7599462479886976e-05, + "loss": 0.7979, + "step": 299 + }, + { + "epoch": 0.9103840682788051, + "grad_norm": 2.118502140045166, + "learning_rate": 1.7576419575412028e-05, + "loss": 0.8007, + "step": 300 + }, + { + "epoch": 0.9103840682788051, + "eval_loss": NaN, + "eval_runtime": 203.6703, + "eval_samples_per_second": 20.415, + "eval_steps_per_second": 5.106, + "step": 300 + }, + { + "epoch": 0.9134186818397345, + "grad_norm": 2.2053334712982178, + "learning_rate": 1.755328184276517e-05, + "loss": 0.7913, + "step": 301 + }, + { + "epoch": 0.9164532954006638, + "grad_norm": 2.243788957595825, + "learning_rate": 1.7530049571543464e-05, + "loss": 0.801, + "step": 302 + }, + { + "epoch": 0.9194879089615932, + "grad_norm": 2.363306760787964, + "learning_rate": 1.7506723052527243e-05, + "loss": 0.8278, + "step": 303 + }, + { + "epoch": 0.9225225225225225, + "grad_norm": 2.2566967010498047, + "learning_rate": 1.7483302577676475e-05, + "loss": 0.7929, + "step": 304 + }, + { + "epoch": 0.9255571360834519, + "grad_norm": 2.1566100120544434, + "learning_rate": 1.7459788440127083e-05, + "loss": 0.7953, + "step": 305 + }, + { + "epoch": 0.9285917496443812, + "grad_norm": 2.279130697250366, + "learning_rate": 1.7436180934187307e-05, + "loss": 0.8125, + "step": 306 + }, + { + "epoch": 0.9316263632053106, + "grad_norm": 2.172891616821289, + "learning_rate": 1.7412480355334006e-05, + "loss": 0.8007, + "step": 307 + }, + { + "epoch": 0.9346609767662399, + "grad_norm": 2.2873098850250244, + "learning_rate": 1.738868700020895e-05, + "loss": 0.779, + "step": 308 + }, + { + "epoch": 0.9376955903271693, + "grad_norm": 2.3155357837677, + "learning_rate": 1.7364801166615124e-05, + "loss": 0.8025, + "step": 309 + }, + { + "epoch": 0.9407302038880986, + "grad_norm": 2.20151686668396, + "learning_rate": 1.7340823153513003e-05, + "loss": 0.8173, + "step": 310 + }, + { + "epoch": 0.943764817449028, + "grad_norm": 2.2033958435058594, + "learning_rate": 1.7316753261016782e-05, + "loss": 0.8095, + "step": 311 + }, + { + "epoch": 0.9467994310099573, + "grad_norm": 2.2805070877075195, + "learning_rate": 1.7292591790390668e-05, + "loss": 0.8139, + "step": 312 + }, + { + "epoch": 0.9498340445708867, + "grad_norm": 2.335238456726074, + "learning_rate": 1.7268339044045044e-05, + "loss": 0.7898, + "step": 313 + }, + { + "epoch": 0.952868658131816, + "grad_norm": 2.1889333724975586, + "learning_rate": 1.7243995325532755e-05, + "loss": 0.8002, + "step": 314 + }, + { + "epoch": 0.9559032716927454, + "grad_norm": 2.5007987022399902, + "learning_rate": 1.7219560939545246e-05, + "loss": 0.7892, + "step": 315 + }, + { + "epoch": 0.9589378852536747, + "grad_norm": 2.1342813968658447, + "learning_rate": 1.7195036191908798e-05, + "loss": 0.8028, + "step": 316 + }, + { + "epoch": 0.9619724988146041, + "grad_norm": 2.199307918548584, + "learning_rate": 1.7170421389580666e-05, + "loss": 0.793, + "step": 317 + }, + { + "epoch": 0.9650071123755334, + "grad_norm": 2.1955904960632324, + "learning_rate": 1.7145716840645253e-05, + "loss": 0.8085, + "step": 318 + }, + { + "epoch": 0.9680417259364628, + "grad_norm": 2.41741943359375, + "learning_rate": 1.712092285431026e-05, + "loss": 0.7964, + "step": 319 + }, + { + "epoch": 0.9710763394973921, + "grad_norm": 2.3595402240753174, + "learning_rate": 1.7096039740902782e-05, + "loss": 0.7999, + "step": 320 + }, + { + "epoch": 0.9741109530583215, + "grad_norm": 2.15049147605896, + "learning_rate": 1.7071067811865477e-05, + "loss": 0.7813, + "step": 321 + }, + { + "epoch": 0.9771455666192508, + "grad_norm": 2.240618944168091, + "learning_rate": 1.7046007379752624e-05, + "loss": 0.8038, + "step": 322 + }, + { + "epoch": 0.9801801801801802, + "grad_norm": 2.418973922729492, + "learning_rate": 1.702085875822623e-05, + "loss": 0.7672, + "step": 323 + }, + { + "epoch": 0.9832147937411095, + "grad_norm": 2.2924294471740723, + "learning_rate": 1.6995622262052093e-05, + "loss": 0.8013, + "step": 324 + }, + { + "epoch": 0.9862494073020389, + "grad_norm": 2.2135136127471924, + "learning_rate": 1.6970298207095887e-05, + "loss": 0.8112, + "step": 325 + }, + { + "epoch": 0.9892840208629682, + "grad_norm": 2.2720751762390137, + "learning_rate": 1.6944886910319173e-05, + "loss": 0.7896, + "step": 326 + }, + { + "epoch": 0.9923186344238976, + "grad_norm": 2.172165632247925, + "learning_rate": 1.6919388689775463e-05, + "loss": 0.8084, + "step": 327 + }, + { + "epoch": 0.9953532479848269, + "grad_norm": 2.1608591079711914, + "learning_rate": 1.6893803864606224e-05, + "loss": 0.7914, + "step": 328 + }, + { + "epoch": 0.9983878615457563, + "grad_norm": 2.153231143951416, + "learning_rate": 1.6868132755036875e-05, + "loss": 0.803, + "step": 329 + }, + { + "epoch": 1.0014224751066856, + "grad_norm": 2.242035150527954, + "learning_rate": 1.6842375682372803e-05, + "loss": 0.7964, + "step": 330 + }, + { + "epoch": 1.004457088667615, + "grad_norm": 2.490118980407715, + "learning_rate": 1.681653296899533e-05, + "loss": 0.7645, + "step": 331 + }, + { + "epoch": 1.0074917022285443, + "grad_norm": 2.258561372756958, + "learning_rate": 1.6790604938357664e-05, + "loss": 0.7579, + "step": 332 + }, + { + "epoch": 1.0105263157894737, + "grad_norm": 2.3750267028808594, + "learning_rate": 1.676459191498087e-05, + "loss": 0.7849, + "step": 333 + }, + { + "epoch": 1.013560929350403, + "grad_norm": 2.5088164806365967, + "learning_rate": 1.6738494224449802e-05, + "loss": 0.7814, + "step": 334 + }, + { + "epoch": 1.0165955429113325, + "grad_norm": 2.4220824241638184, + "learning_rate": 1.6712312193409032e-05, + "loss": 0.7774, + "step": 335 + }, + { + "epoch": 1.0196301564722616, + "grad_norm": 2.2030773162841797, + "learning_rate": 1.6686046149558736e-05, + "loss": 0.7495, + "step": 336 + }, + { + "epoch": 1.022664770033191, + "grad_norm": 2.336583375930786, + "learning_rate": 1.6659696421650645e-05, + "loss": 0.7508, + "step": 337 + }, + { + "epoch": 1.0256993835941204, + "grad_norm": 2.3112359046936035, + "learning_rate": 1.6633263339483867e-05, + "loss": 0.7525, + "step": 338 + }, + { + "epoch": 1.0287339971550498, + "grad_norm": 2.165022611618042, + "learning_rate": 1.6606747233900816e-05, + "loss": 0.787, + "step": 339 + }, + { + "epoch": 1.0317686107159791, + "grad_norm": 2.3545596599578857, + "learning_rate": 1.658014843678303e-05, + "loss": 0.7665, + "step": 340 + }, + { + "epoch": 1.0348032242769085, + "grad_norm": 2.231351137161255, + "learning_rate": 1.655346728104704e-05, + "loss": 0.7726, + "step": 341 + }, + { + "epoch": 1.037837837837838, + "grad_norm": 2.292428970336914, + "learning_rate": 1.652670410064019e-05, + "loss": 0.7722, + "step": 342 + }, + { + "epoch": 1.0408724513987673, + "grad_norm": 2.2555713653564453, + "learning_rate": 1.6499859230536468e-05, + "loss": 0.755, + "step": 343 + }, + { + "epoch": 1.0439070649596967, + "grad_norm": 2.5129449367523193, + "learning_rate": 1.647293300673231e-05, + "loss": 0.7736, + "step": 344 + }, + { + "epoch": 1.0469416785206258, + "grad_norm": 2.269122362136841, + "learning_rate": 1.6445925766242392e-05, + "loss": 0.7898, + "step": 345 + }, + { + "epoch": 1.0499762920815552, + "grad_norm": 2.21991229057312, + "learning_rate": 1.641883784709541e-05, + "loss": 0.7767, + "step": 346 + }, + { + "epoch": 1.0530109056424846, + "grad_norm": 2.324253797531128, + "learning_rate": 1.639166958832985e-05, + "loss": 0.7728, + "step": 347 + }, + { + "epoch": 1.056045519203414, + "grad_norm": 2.3205628395080566, + "learning_rate": 1.6364421329989758e-05, + "loss": 0.7845, + "step": 348 + }, + { + "epoch": 1.0590801327643433, + "grad_norm": 2.361678123474121, + "learning_rate": 1.6337093413120463e-05, + "loss": 0.7455, + "step": 349 + }, + { + "epoch": 1.0621147463252727, + "grad_norm": 2.3375606536865234, + "learning_rate": 1.6309686179764317e-05, + "loss": 0.7754, + "step": 350 + }, + { + "epoch": 1.0621147463252727, + "eval_loss": NaN, + "eval_runtime": 203.6306, + "eval_samples_per_second": 20.419, + "eval_steps_per_second": 5.107, + "step": 350 + }, + { + "epoch": 1.065149359886202, + "grad_norm": 2.3522422313690186, + "learning_rate": 1.6282199972956425e-05, + "loss": 0.7759, + "step": 351 + }, + { + "epoch": 1.0681839734471312, + "grad_norm": 2.4227213859558105, + "learning_rate": 1.6254635136720328e-05, + "loss": 0.7772, + "step": 352 + }, + { + "epoch": 1.0712185870080606, + "grad_norm": 2.235722303390503, + "learning_rate": 1.6226992016063726e-05, + "loss": 0.7694, + "step": 353 + }, + { + "epoch": 1.07425320056899, + "grad_norm": 2.2417314052581787, + "learning_rate": 1.6199270956974128e-05, + "loss": 0.7628, + "step": 354 + }, + { + "epoch": 1.0772878141299194, + "grad_norm": 2.348954200744629, + "learning_rate": 1.6171472306414554e-05, + "loss": 0.7656, + "step": 355 + }, + { + "epoch": 1.0803224276908487, + "grad_norm": 2.346963882446289, + "learning_rate": 1.614359641231916e-05, + "loss": 0.7839, + "step": 356 + }, + { + "epoch": 1.0833570412517781, + "grad_norm": 2.2969138622283936, + "learning_rate": 1.6115643623588915e-05, + "loss": 0.7728, + "step": 357 + }, + { + "epoch": 1.0863916548127075, + "grad_norm": 2.3338327407836914, + "learning_rate": 1.608761429008721e-05, + "loss": 0.7902, + "step": 358 + }, + { + "epoch": 1.0894262683736369, + "grad_norm": 2.2462401390075684, + "learning_rate": 1.6059508762635482e-05, + "loss": 0.7906, + "step": 359 + }, + { + "epoch": 1.0924608819345663, + "grad_norm": 2.2091758251190186, + "learning_rate": 1.6031327393008848e-05, + "loss": 0.7587, + "step": 360 + }, + { + "epoch": 1.0954954954954954, + "grad_norm": 2.2392489910125732, + "learning_rate": 1.6003070533931657e-05, + "loss": 0.7598, + "step": 361 + }, + { + "epoch": 1.0985301090564248, + "grad_norm": 2.193833112716675, + "learning_rate": 1.5974738539073125e-05, + "loss": 0.7622, + "step": 362 + }, + { + "epoch": 1.1015647226173542, + "grad_norm": 2.187610149383545, + "learning_rate": 1.594633176304287e-05, + "loss": 0.7796, + "step": 363 + }, + { + "epoch": 1.1045993361782835, + "grad_norm": 2.2753069400787354, + "learning_rate": 1.5917850561386487e-05, + "loss": 0.7783, + "step": 364 + }, + { + "epoch": 1.107633949739213, + "grad_norm": 2.2835614681243896, + "learning_rate": 1.588929529058111e-05, + "loss": 0.7801, + "step": 365 + }, + { + "epoch": 1.1106685633001423, + "grad_norm": 2.46498441696167, + "learning_rate": 1.5860666308030933e-05, + "loss": 0.7683, + "step": 366 + }, + { + "epoch": 1.1137031768610717, + "grad_norm": 2.361351490020752, + "learning_rate": 1.5831963972062734e-05, + "loss": 0.783, + "step": 367 + }, + { + "epoch": 1.1167377904220008, + "grad_norm": 2.2396347522735596, + "learning_rate": 1.5803188641921417e-05, + "loss": 0.7563, + "step": 368 + }, + { + "epoch": 1.1197724039829302, + "grad_norm": 2.2810609340667725, + "learning_rate": 1.5774340677765483e-05, + "loss": 0.7865, + "step": 369 + }, + { + "epoch": 1.1228070175438596, + "grad_norm": 2.147937297821045, + "learning_rate": 1.5745420440662543e-05, + "loss": 0.7684, + "step": 370 + }, + { + "epoch": 1.125841631104789, + "grad_norm": 2.355337142944336, + "learning_rate": 1.5716428292584788e-05, + "loss": 0.7693, + "step": 371 + }, + { + "epoch": 1.1288762446657183, + "grad_norm": 2.4234957695007324, + "learning_rate": 1.568736459640447e-05, + "loss": 0.7567, + "step": 372 + }, + { + "epoch": 1.1319108582266477, + "grad_norm": 2.2321126461029053, + "learning_rate": 1.5658229715889345e-05, + "loss": 0.7984, + "step": 373 + }, + { + "epoch": 1.134945471787577, + "grad_norm": 2.2270772457122803, + "learning_rate": 1.5629024015698137e-05, + "loss": 0.7868, + "step": 374 + }, + { + "epoch": 1.1379800853485065, + "grad_norm": 2.4906022548675537, + "learning_rate": 1.5599747861375957e-05, + "loss": 0.7761, + "step": 375 + }, + { + "epoch": 1.1410146989094359, + "grad_norm": 2.4099533557891846, + "learning_rate": 1.5570401619349737e-05, + "loss": 0.7727, + "step": 376 + }, + { + "epoch": 1.144049312470365, + "grad_norm": 2.167451858520508, + "learning_rate": 1.5540985656923648e-05, + "loss": 0.7449, + "step": 377 + }, + { + "epoch": 1.1470839260312944, + "grad_norm": 2.4137990474700928, + "learning_rate": 1.551150034227449e-05, + "loss": 0.7836, + "step": 378 + }, + { + "epoch": 1.1501185395922238, + "grad_norm": 2.0170676708221436, + "learning_rate": 1.54819460444471e-05, + "loss": 0.771, + "step": 379 + }, + { + "epoch": 1.1531531531531531, + "grad_norm": 2.3556909561157227, + "learning_rate": 1.5452323133349712e-05, + "loss": 0.769, + "step": 380 + }, + { + "epoch": 1.1561877667140825, + "grad_norm": 2.1490001678466797, + "learning_rate": 1.5422631979749354e-05, + "loss": 0.7467, + "step": 381 + }, + { + "epoch": 1.159222380275012, + "grad_norm": 2.240185260772705, + "learning_rate": 1.5392872955267176e-05, + "loss": 0.7677, + "step": 382 + }, + { + "epoch": 1.1622569938359413, + "grad_norm": 2.1402430534362793, + "learning_rate": 1.5363046432373824e-05, + "loss": 0.7706, + "step": 383 + }, + { + "epoch": 1.1652916073968704, + "grad_norm": 2.043536424636841, + "learning_rate": 1.5333152784384777e-05, + "loss": 0.7679, + "step": 384 + }, + { + "epoch": 1.1683262209577998, + "grad_norm": 2.146432399749756, + "learning_rate": 1.5303192385455652e-05, + "loss": 0.7746, + "step": 385 + }, + { + "epoch": 1.1713608345187292, + "grad_norm": 2.216525077819824, + "learning_rate": 1.5273165610577543e-05, + "loss": 0.7735, + "step": 386 + }, + { + "epoch": 1.1743954480796586, + "grad_norm": 2.134411334991455, + "learning_rate": 1.5243072835572319e-05, + "loss": 0.771, + "step": 387 + }, + { + "epoch": 1.177430061640588, + "grad_norm": 2.210275173187256, + "learning_rate": 1.5212914437087921e-05, + "loss": 0.7665, + "step": 388 + }, + { + "epoch": 1.1804646752015173, + "grad_norm": 2.2126176357269287, + "learning_rate": 1.5182690792593659e-05, + "loss": 0.7658, + "step": 389 + }, + { + "epoch": 1.1834992887624467, + "grad_norm": 1.9885146617889404, + "learning_rate": 1.5152402280375454e-05, + "loss": 0.7509, + "step": 390 + }, + { + "epoch": 1.186533902323376, + "grad_norm": 2.2027952671051025, + "learning_rate": 1.5122049279531143e-05, + "loss": 0.7811, + "step": 391 + }, + { + "epoch": 1.1895685158843055, + "grad_norm": 2.1936960220336914, + "learning_rate": 1.509163216996572e-05, + "loss": 0.7785, + "step": 392 + }, + { + "epoch": 1.1926031294452346, + "grad_norm": 2.166518449783325, + "learning_rate": 1.5061151332386565e-05, + "loss": 0.7775, + "step": 393 + }, + { + "epoch": 1.195637743006164, + "grad_norm": 2.0909955501556396, + "learning_rate": 1.5030607148298697e-05, + "loss": 0.7783, + "step": 394 + }, + { + "epoch": 1.1986723565670934, + "grad_norm": 2.287322521209717, + "learning_rate": 1.5000000000000002e-05, + "loss": 0.7678, + "step": 395 + }, + { + "epoch": 1.2017069701280227, + "grad_norm": 2.265106678009033, + "learning_rate": 1.4969330270576428e-05, + "loss": 0.7772, + "step": 396 + }, + { + "epoch": 1.2047415836889521, + "grad_norm": 2.2475264072418213, + "learning_rate": 1.4938598343897215e-05, + "loss": 0.7509, + "step": 397 + }, + { + "epoch": 1.2077761972498815, + "grad_norm": 2.3240981101989746, + "learning_rate": 1.4907804604610064e-05, + "loss": 0.7849, + "step": 398 + }, + { + "epoch": 1.2108108108108109, + "grad_norm": 2.148869276046753, + "learning_rate": 1.4876949438136348e-05, + "loss": 0.7781, + "step": 399 + }, + { + "epoch": 1.2138454243717403, + "grad_norm": 2.318875789642334, + "learning_rate": 1.484603323066627e-05, + "loss": 0.7987, + "step": 400 + }, + { + "epoch": 1.2138454243717403, + "eval_loss": NaN, + "eval_runtime": 204.8025, + "eval_samples_per_second": 20.302, + "eval_steps_per_second": 5.078, + "step": 400 + }, + { + "epoch": 1.2168800379326696, + "grad_norm": 2.1469035148620605, + "learning_rate": 1.4815056369154039e-05, + "loss": 0.7924, + "step": 401 + }, + { + "epoch": 1.2199146514935988, + "grad_norm": 2.3983654975891113, + "learning_rate": 1.4784019241313025e-05, + "loss": 0.7431, + "step": 402 + }, + { + "epoch": 1.2229492650545282, + "grad_norm": 2.1171765327453613, + "learning_rate": 1.47529222356109e-05, + "loss": 0.7583, + "step": 403 + }, + { + "epoch": 1.2259838786154575, + "grad_norm": 2.3186557292938232, + "learning_rate": 1.4721765741264786e-05, + "loss": 0.7545, + "step": 404 + }, + { + "epoch": 1.229018492176387, + "grad_norm": 2.308945417404175, + "learning_rate": 1.4690550148236371e-05, + "loss": 0.7752, + "step": 405 + }, + { + "epoch": 1.2320531057373163, + "grad_norm": 2.141418933868408, + "learning_rate": 1.4659275847227044e-05, + "loss": 0.7501, + "step": 406 + }, + { + "epoch": 1.2350877192982457, + "grad_norm": 2.1447696685791016, + "learning_rate": 1.4627943229672992e-05, + "loss": 0.7446, + "step": 407 + }, + { + "epoch": 1.238122332859175, + "grad_norm": 2.062683582305908, + "learning_rate": 1.4596552687740304e-05, + "loss": 0.7729, + "step": 408 + }, + { + "epoch": 1.2411569464201042, + "grad_norm": 2.283247232437134, + "learning_rate": 1.4565104614320065e-05, + "loss": 0.7752, + "step": 409 + }, + { + "epoch": 1.2441915599810336, + "grad_norm": 2.299151659011841, + "learning_rate": 1.453359940302344e-05, + "loss": 0.7794, + "step": 410 + }, + { + "epoch": 1.247226173541963, + "grad_norm": 2.2340760231018066, + "learning_rate": 1.4502037448176734e-05, + "loss": 0.7811, + "step": 411 + }, + { + "epoch": 1.2502607871028923, + "grad_norm": 2.305233955383301, + "learning_rate": 1.4470419144816483e-05, + "loss": 0.7663, + "step": 412 + }, + { + "epoch": 1.2532954006638217, + "grad_norm": 2.1460888385772705, + "learning_rate": 1.4438744888684481e-05, + "loss": 0.7584, + "step": 413 + }, + { + "epoch": 1.256330014224751, + "grad_norm": 2.3122851848602295, + "learning_rate": 1.4407015076222845e-05, + "loss": 0.7817, + "step": 414 + }, + { + "epoch": 1.2593646277856805, + "grad_norm": 2.0583643913269043, + "learning_rate": 1.4375230104569044e-05, + "loss": 0.7695, + "step": 415 + }, + { + "epoch": 1.2623992413466096, + "grad_norm": 2.262274980545044, + "learning_rate": 1.4343390371550936e-05, + "loss": 0.7739, + "step": 416 + }, + { + "epoch": 1.2654338549075392, + "grad_norm": 2.151893377304077, + "learning_rate": 1.4311496275681785e-05, + "loss": 0.7789, + "step": 417 + }, + { + "epoch": 1.2684684684684684, + "grad_norm": 2.28367280960083, + "learning_rate": 1.4279548216155265e-05, + "loss": 0.775, + "step": 418 + }, + { + "epoch": 1.2715030820293978, + "grad_norm": 2.3059751987457275, + "learning_rate": 1.424754659284048e-05, + "loss": 0.7613, + "step": 419 + }, + { + "epoch": 1.2745376955903271, + "grad_norm": 2.436896800994873, + "learning_rate": 1.4215491806276944e-05, + "loss": 0.7835, + "step": 420 + }, + { + "epoch": 1.2775723091512565, + "grad_norm": 2.102220058441162, + "learning_rate": 1.418338425766958e-05, + "loss": 0.7932, + "step": 421 + }, + { + "epoch": 1.280606922712186, + "grad_norm": 2.1711723804473877, + "learning_rate": 1.4151224348883692e-05, + "loss": 0.7668, + "step": 422 + }, + { + "epoch": 1.2836415362731153, + "grad_norm": 2.213289737701416, + "learning_rate": 1.4119012482439929e-05, + "loss": 0.7745, + "step": 423 + }, + { + "epoch": 1.2866761498340447, + "grad_norm": 2.0960137844085693, + "learning_rate": 1.408674906150926e-05, + "loss": 0.7742, + "step": 424 + }, + { + "epoch": 1.2897107633949738, + "grad_norm": 2.081200122833252, + "learning_rate": 1.4054434489907916e-05, + "loss": 0.7652, + "step": 425 + }, + { + "epoch": 1.2927453769559032, + "grad_norm": 2.284423351287842, + "learning_rate": 1.4022069172092354e-05, + "loss": 0.7762, + "step": 426 + }, + { + "epoch": 1.2957799905168326, + "grad_norm": 2.292639970779419, + "learning_rate": 1.3989653513154165e-05, + "loss": 0.7644, + "step": 427 + }, + { + "epoch": 1.298814604077762, + "grad_norm": 2.282759428024292, + "learning_rate": 1.3957187918815032e-05, + "loss": 0.7658, + "step": 428 + }, + { + "epoch": 1.3018492176386913, + "grad_norm": 2.2170190811157227, + "learning_rate": 1.3924672795421638e-05, + "loss": 0.7661, + "step": 429 + }, + { + "epoch": 1.3048838311996207, + "grad_norm": 2.202991247177124, + "learning_rate": 1.3892108549940583e-05, + "loss": 0.7881, + "step": 430 + }, + { + "epoch": 1.30791844476055, + "grad_norm": 2.148986339569092, + "learning_rate": 1.3859495589953289e-05, + "loss": 0.7865, + "step": 431 + }, + { + "epoch": 1.3109530583214792, + "grad_norm": 2.1265058517456055, + "learning_rate": 1.3826834323650899e-05, + "loss": 0.789, + "step": 432 + }, + { + "epoch": 1.3139876718824088, + "grad_norm": 2.1798737049102783, + "learning_rate": 1.3794125159829173e-05, + "loss": 0.7707, + "step": 433 + }, + { + "epoch": 1.317022285443338, + "grad_norm": 2.2072978019714355, + "learning_rate": 1.376136850788336e-05, + "loss": 0.763, + "step": 434 + }, + { + "epoch": 1.3200568990042674, + "grad_norm": 2.12349534034729, + "learning_rate": 1.3728564777803089e-05, + "loss": 0.7505, + "step": 435 + }, + { + "epoch": 1.3230915125651967, + "grad_norm": 2.104276180267334, + "learning_rate": 1.3695714380167221e-05, + "loss": 0.7891, + "step": 436 + }, + { + "epoch": 1.3261261261261261, + "grad_norm": 2.038515090942383, + "learning_rate": 1.3662817726138729e-05, + "loss": 0.7668, + "step": 437 + }, + { + "epoch": 1.3291607396870555, + "grad_norm": 2.0557668209075928, + "learning_rate": 1.3629875227459532e-05, + "loss": 0.7685, + "step": 438 + }, + { + "epoch": 1.3321953532479849, + "grad_norm": 2.221299886703491, + "learning_rate": 1.359688729644536e-05, + "loss": 0.7765, + "step": 439 + }, + { + "epoch": 1.3352299668089143, + "grad_norm": 2.383873462677002, + "learning_rate": 1.356385434598057e-05, + "loss": 0.7863, + "step": 440 + }, + { + "epoch": 1.3382645803698434, + "grad_norm": 2.144969940185547, + "learning_rate": 1.3530776789513009e-05, + "loss": 0.7854, + "step": 441 + }, + { + "epoch": 1.341299193930773, + "grad_norm": 2.3431177139282227, + "learning_rate": 1.3497655041048812e-05, + "loss": 0.7491, + "step": 442 + }, + { + "epoch": 1.3443338074917022, + "grad_norm": 2.1558756828308105, + "learning_rate": 1.3464489515147239e-05, + "loss": 0.7935, + "step": 443 + }, + { + "epoch": 1.3473684210526315, + "grad_norm": 2.4032328128814697, + "learning_rate": 1.3431280626915466e-05, + "loss": 0.765, + "step": 444 + }, + { + "epoch": 1.350403034613561, + "grad_norm": 2.373549461364746, + "learning_rate": 1.3398028792003413e-05, + "loss": 0.766, + "step": 445 + }, + { + "epoch": 1.3534376481744903, + "grad_norm": 2.38440203666687, + "learning_rate": 1.3364734426598527e-05, + "loss": 0.7849, + "step": 446 + }, + { + "epoch": 1.3564722617354197, + "grad_norm": 2.319101333618164, + "learning_rate": 1.3331397947420578e-05, + "loss": 0.7738, + "step": 447 + }, + { + "epoch": 1.359506875296349, + "grad_norm": 2.1911866664886475, + "learning_rate": 1.3298019771716435e-05, + "loss": 0.7779, + "step": 448 + }, + { + "epoch": 1.3625414888572784, + "grad_norm": 2.273451328277588, + "learning_rate": 1.3264600317254854e-05, + "loss": 0.76, + "step": 449 + }, + { + "epoch": 1.3655761024182076, + "grad_norm": 2.2576355934143066, + "learning_rate": 1.3231140002321252e-05, + "loss": 0.7687, + "step": 450 + }, + { + "epoch": 1.3655761024182076, + "eval_loss": NaN, + "eval_runtime": 203.9814, + "eval_samples_per_second": 20.384, + "eval_steps_per_second": 5.099, + "step": 450 + }, + { + "epoch": 1.368610715979137, + "grad_norm": 2.2846243381500244, + "learning_rate": 1.3197639245712454e-05, + "loss": 0.77, + "step": 451 + }, + { + "epoch": 1.3716453295400663, + "grad_norm": 2.1583635807037354, + "learning_rate": 1.3164098466731467e-05, + "loss": 0.7681, + "step": 452 + }, + { + "epoch": 1.3746799431009957, + "grad_norm": 2.126995325088501, + "learning_rate": 1.3130518085182224e-05, + "loss": 0.7755, + "step": 453 + }, + { + "epoch": 1.377714556661925, + "grad_norm": 2.2357747554779053, + "learning_rate": 1.3096898521364338e-05, + "loss": 0.7509, + "step": 454 + }, + { + "epoch": 1.3807491702228545, + "grad_norm": 2.0657386779785156, + "learning_rate": 1.3063240196067837e-05, + "loss": 0.8043, + "step": 455 + }, + { + "epoch": 1.3837837837837839, + "grad_norm": 2.1653802394866943, + "learning_rate": 1.3029543530567884e-05, + "loss": 0.7676, + "step": 456 + }, + { + "epoch": 1.386818397344713, + "grad_norm": 2.161508798599243, + "learning_rate": 1.2995808946619533e-05, + "loss": 0.7735, + "step": 457 + }, + { + "epoch": 1.3898530109056426, + "grad_norm": 2.185350179672241, + "learning_rate": 1.2962036866452423e-05, + "loss": 0.7891, + "step": 458 + }, + { + "epoch": 1.3928876244665718, + "grad_norm": 2.198807954788208, + "learning_rate": 1.2928227712765504e-05, + "loss": 0.7657, + "step": 459 + }, + { + "epoch": 1.3959222380275011, + "grad_norm": 2.2574117183685303, + "learning_rate": 1.2894381908721757e-05, + "loss": 0.7893, + "step": 460 + }, + { + "epoch": 1.3989568515884305, + "grad_norm": 2.2577221393585205, + "learning_rate": 1.2860499877942876e-05, + "loss": 0.7787, + "step": 461 + }, + { + "epoch": 1.40199146514936, + "grad_norm": 2.265421152114868, + "learning_rate": 1.282658204450398e-05, + "loss": 0.7714, + "step": 462 + }, + { + "epoch": 1.4050260787102893, + "grad_norm": 2.0731489658355713, + "learning_rate": 1.2792628832928302e-05, + "loss": 0.782, + "step": 463 + }, + { + "epoch": 1.4080606922712187, + "grad_norm": 2.130533218383789, + "learning_rate": 1.275864066818188e-05, + "loss": 0.7873, + "step": 464 + }, + { + "epoch": 1.411095305832148, + "grad_norm": 2.0858917236328125, + "learning_rate": 1.2724617975668229e-05, + "loss": 0.8005, + "step": 465 + }, + { + "epoch": 1.4141299193930772, + "grad_norm": 2.0890703201293945, + "learning_rate": 1.2690561181223024e-05, + "loss": 0.775, + "step": 466 + }, + { + "epoch": 1.4171645329540066, + "grad_norm": 2.3934412002563477, + "learning_rate": 1.2656470711108763e-05, + "loss": 0.7931, + "step": 467 + }, + { + "epoch": 1.420199146514936, + "grad_norm": 2.2045512199401855, + "learning_rate": 1.2622346992009447e-05, + "loss": 0.7747, + "step": 468 + }, + { + "epoch": 1.4232337600758653, + "grad_norm": 2.371346950531006, + "learning_rate": 1.2588190451025209e-05, + "loss": 0.7581, + "step": 469 + }, + { + "epoch": 1.4262683736367947, + "grad_norm": 2.1523914337158203, + "learning_rate": 1.2554001515667009e-05, + "loss": 0.7741, + "step": 470 + }, + { + "epoch": 1.429302987197724, + "grad_norm": 2.2948853969573975, + "learning_rate": 1.2519780613851254e-05, + "loss": 0.7925, + "step": 471 + }, + { + "epoch": 1.4323376007586535, + "grad_norm": 2.2751922607421875, + "learning_rate": 1.2485528173894447e-05, + "loss": 0.7784, + "step": 472 + }, + { + "epoch": 1.4353722143195826, + "grad_norm": 2.1596484184265137, + "learning_rate": 1.2451244624507831e-05, + "loss": 0.7895, + "step": 473 + }, + { + "epoch": 1.4384068278805122, + "grad_norm": 2.2292542457580566, + "learning_rate": 1.2416930394792026e-05, + "loss": 0.7698, + "step": 474 + }, + { + "epoch": 1.4414414414414414, + "grad_norm": 2.1559245586395264, + "learning_rate": 1.238258591423165e-05, + "loss": 0.776, + "step": 475 + }, + { + "epoch": 1.4444760550023708, + "grad_norm": 2.072768211364746, + "learning_rate": 1.234821161268995e-05, + "loss": 0.7665, + "step": 476 + }, + { + "epoch": 1.4475106685633001, + "grad_norm": 2.3749032020568848, + "learning_rate": 1.2313807920403419e-05, + "loss": 0.7765, + "step": 477 + }, + { + "epoch": 1.4505452821242295, + "grad_norm": 2.1614534854888916, + "learning_rate": 1.22793752679764e-05, + "loss": 0.7908, + "step": 478 + }, + { + "epoch": 1.4535798956851589, + "grad_norm": 2.2183918952941895, + "learning_rate": 1.2244914086375726e-05, + "loss": 0.7662, + "step": 479 + }, + { + "epoch": 1.4566145092460883, + "grad_norm": 2.230243444442749, + "learning_rate": 1.22104248069253e-05, + "loss": 0.7758, + "step": 480 + }, + { + "epoch": 1.4596491228070176, + "grad_norm": 2.503591775894165, + "learning_rate": 1.2175907861300698e-05, + "loss": 0.7739, + "step": 481 + }, + { + "epoch": 1.4626837363679468, + "grad_norm": 2.4481165409088135, + "learning_rate": 1.2141363681523777e-05, + "loss": 0.788, + "step": 482 + }, + { + "epoch": 1.4657183499288762, + "grad_norm": 2.30068302154541, + "learning_rate": 1.2106792699957264e-05, + "loss": 0.7905, + "step": 483 + }, + { + "epoch": 1.4687529634898056, + "grad_norm": 2.2382349967956543, + "learning_rate": 1.2072195349299344e-05, + "loss": 0.7617, + "step": 484 + }, + { + "epoch": 1.471787577050735, + "grad_norm": 2.2054882049560547, + "learning_rate": 1.2037572062578238e-05, + "loss": 0.7802, + "step": 485 + }, + { + "epoch": 1.4748221906116643, + "grad_norm": 2.282318115234375, + "learning_rate": 1.2002923273146793e-05, + "loss": 0.7711, + "step": 486 + }, + { + "epoch": 1.4778568041725937, + "grad_norm": 2.1077611446380615, + "learning_rate": 1.1968249414677055e-05, + "loss": 0.761, + "step": 487 + }, + { + "epoch": 1.480891417733523, + "grad_norm": 2.1558871269226074, + "learning_rate": 1.1933550921154836e-05, + "loss": 0.7872, + "step": 488 + }, + { + "epoch": 1.4839260312944522, + "grad_norm": 2.332897901535034, + "learning_rate": 1.1898828226874284e-05, + "loss": 0.7791, + "step": 489 + }, + { + "epoch": 1.4869606448553818, + "grad_norm": 2.0427420139312744, + "learning_rate": 1.1864081766432457e-05, + "loss": 0.782, + "step": 490 + }, + { + "epoch": 1.489995258416311, + "grad_norm": 2.2689926624298096, + "learning_rate": 1.1829311974723868e-05, + "loss": 0.7659, + "step": 491 + }, + { + "epoch": 1.4930298719772404, + "grad_norm": 2.2432594299316406, + "learning_rate": 1.1794519286935056e-05, + "loss": 0.7543, + "step": 492 + }, + { + "epoch": 1.4960644855381697, + "grad_norm": 2.0946452617645264, + "learning_rate": 1.1759704138539121e-05, + "loss": 0.7712, + "step": 493 + }, + { + "epoch": 1.499099099099099, + "grad_norm": 2.1154541969299316, + "learning_rate": 1.1724866965290302e-05, + "loss": 0.7732, + "step": 494 + }, + { + "epoch": 1.5021337126600285, + "grad_norm": 2.195223569869995, + "learning_rate": 1.1690008203218493e-05, + "loss": 0.7596, + "step": 495 + }, + { + "epoch": 1.5051683262209576, + "grad_norm": 2.105151653289795, + "learning_rate": 1.1655128288623803e-05, + "loss": 0.775, + "step": 496 + }, + { + "epoch": 1.5082029397818872, + "grad_norm": 2.196159601211548, + "learning_rate": 1.1620227658071088e-05, + "loss": 0.7893, + "step": 497 + }, + { + "epoch": 1.5112375533428164, + "grad_norm": 2.3409769535064697, + "learning_rate": 1.158530674838449e-05, + "loss": 0.7494, + "step": 498 + }, + { + "epoch": 1.514272166903746, + "grad_norm": 2.177128553390503, + "learning_rate": 1.155036599664198e-05, + "loss": 0.7595, + "step": 499 + }, + { + "epoch": 1.5173067804646752, + "grad_norm": 2.3704683780670166, + "learning_rate": 1.1515405840169861e-05, + "loss": 0.7607, + "step": 500 + }, + { + "epoch": 1.5173067804646752, + "eval_loss": NaN, + "eval_runtime": 204.5697, + "eval_samples_per_second": 20.326, + "eval_steps_per_second": 5.084, + "step": 500 + }, + { + "epoch": 1.5203413940256045, + "grad_norm": 2.175325393676758, + "learning_rate": 1.1480426716537316e-05, + "loss": 0.7626, + "step": 501 + }, + { + "epoch": 1.523376007586534, + "grad_norm": 2.093395471572876, + "learning_rate": 1.1445429063550925e-05, + "loss": 0.7787, + "step": 502 + }, + { + "epoch": 1.5264106211474633, + "grad_norm": 2.2766175270080566, + "learning_rate": 1.1410413319249193e-05, + "loss": 0.7592, + "step": 503 + }, + { + "epoch": 1.5294452347083927, + "grad_norm": 2.00833797454834, + "learning_rate": 1.1375379921897052e-05, + "loss": 0.7685, + "step": 504 + }, + { + "epoch": 1.5324798482693218, + "grad_norm": 2.2018191814422607, + "learning_rate": 1.1340329309980379e-05, + "loss": 0.753, + "step": 505 + }, + { + "epoch": 1.5355144618302514, + "grad_norm": 2.228724479675293, + "learning_rate": 1.130526192220052e-05, + "loss": 0.7687, + "step": 506 + }, + { + "epoch": 1.5385490753911806, + "grad_norm": 2.241725444793701, + "learning_rate": 1.1270178197468788e-05, + "loss": 0.7674, + "step": 507 + }, + { + "epoch": 1.54158368895211, + "grad_norm": 2.1697564125061035, + "learning_rate": 1.1235078574900984e-05, + "loss": 0.7726, + "step": 508 + }, + { + "epoch": 1.5446183025130393, + "grad_norm": 2.197449207305908, + "learning_rate": 1.119996349381187e-05, + "loss": 0.7672, + "step": 509 + }, + { + "epoch": 1.5476529160739687, + "grad_norm": 2.0595834255218506, + "learning_rate": 1.1164833393709707e-05, + "loss": 0.7706, + "step": 510 + }, + { + "epoch": 1.550687529634898, + "grad_norm": 2.3002591133117676, + "learning_rate": 1.112968871429073e-05, + "loss": 0.7875, + "step": 511 + }, + { + "epoch": 1.5537221431958272, + "grad_norm": 2.175219774246216, + "learning_rate": 1.1094529895433653e-05, + "loss": 0.7809, + "step": 512 + }, + { + "epoch": 1.5567567567567568, + "grad_norm": 2.1368846893310547, + "learning_rate": 1.1059357377194161e-05, + "loss": 0.7878, + "step": 513 + }, + { + "epoch": 1.559791370317686, + "grad_norm": 2.210344076156616, + "learning_rate": 1.102417159979941e-05, + "loss": 0.7543, + "step": 514 + }, + { + "epoch": 1.5628259838786156, + "grad_norm": 2.1888577938079834, + "learning_rate": 1.09889730036425e-05, + "loss": 0.7731, + "step": 515 + }, + { + "epoch": 1.5658605974395448, + "grad_norm": 2.3701112270355225, + "learning_rate": 1.0953762029276982e-05, + "loss": 0.7689, + "step": 516 + }, + { + "epoch": 1.5688952110004741, + "grad_norm": 2.069556713104248, + "learning_rate": 1.0918539117411334e-05, + "loss": 0.767, + "step": 517 + }, + { + "epoch": 1.5719298245614035, + "grad_norm": 2.209773302078247, + "learning_rate": 1.0883304708903441e-05, + "loss": 0.7696, + "step": 518 + }, + { + "epoch": 1.5749644381223329, + "grad_norm": 2.1156795024871826, + "learning_rate": 1.0848059244755093e-05, + "loss": 0.7827, + "step": 519 + }, + { + "epoch": 1.5779990516832623, + "grad_norm": 2.359513521194458, + "learning_rate": 1.0812803166106445e-05, + "loss": 0.7612, + "step": 520 + }, + { + "epoch": 1.5810336652441914, + "grad_norm": 2.2499759197235107, + "learning_rate": 1.0777536914230509e-05, + "loss": 0.7671, + "step": 521 + }, + { + "epoch": 1.584068278805121, + "grad_norm": 2.219525098800659, + "learning_rate": 1.0742260930527625e-05, + "loss": 0.776, + "step": 522 + }, + { + "epoch": 1.5871028923660502, + "grad_norm": 2.339210033416748, + "learning_rate": 1.0706975656519946e-05, + "loss": 0.7669, + "step": 523 + }, + { + "epoch": 1.5901375059269798, + "grad_norm": 2.3282480239868164, + "learning_rate": 1.06716815338459e-05, + "loss": 0.7843, + "step": 524 + }, + { + "epoch": 1.593172119487909, + "grad_norm": 2.09635329246521, + "learning_rate": 1.0636379004254665e-05, + "loss": 0.7598, + "step": 525 + }, + { + "epoch": 1.5962067330488383, + "grad_norm": 2.3128199577331543, + "learning_rate": 1.0601068509600642e-05, + "loss": 0.7673, + "step": 526 + }, + { + "epoch": 1.5992413466097677, + "grad_norm": 2.3669700622558594, + "learning_rate": 1.0565750491837925e-05, + "loss": 0.7697, + "step": 527 + }, + { + "epoch": 1.602275960170697, + "grad_norm": 2.2540953159332275, + "learning_rate": 1.0530425393014773e-05, + "loss": 0.7641, + "step": 528 + }, + { + "epoch": 1.6053105737316264, + "grad_norm": 2.3089439868927, + "learning_rate": 1.049509365526807e-05, + "loss": 0.768, + "step": 529 + }, + { + "epoch": 1.6083451872925556, + "grad_norm": 2.0084259510040283, + "learning_rate": 1.0459755720817797e-05, + "loss": 0.7504, + "step": 530 + }, + { + "epoch": 1.6113798008534852, + "grad_norm": 2.261798143386841, + "learning_rate": 1.0424412031961485e-05, + "loss": 0.7629, + "step": 531 + }, + { + "epoch": 1.6144144144144144, + "grad_norm": 2.2108161449432373, + "learning_rate": 1.0389063031068698e-05, + "loss": 0.738, + "step": 532 + }, + { + "epoch": 1.6174490279753437, + "grad_norm": 2.2087652683258057, + "learning_rate": 1.0353709160575488e-05, + "loss": 0.7758, + "step": 533 + }, + { + "epoch": 1.6204836415362731, + "grad_norm": 2.218838691711426, + "learning_rate": 1.0318350862978848e-05, + "loss": 0.7641, + "step": 534 + }, + { + "epoch": 1.6235182550972025, + "grad_norm": 2.156991720199585, + "learning_rate": 1.0282988580831183e-05, + "loss": 0.7577, + "step": 535 + }, + { + "epoch": 1.6265528686581319, + "grad_norm": 2.1950886249542236, + "learning_rate": 1.0247622756734775e-05, + "loss": 0.7888, + "step": 536 + }, + { + "epoch": 1.629587482219061, + "grad_norm": 2.114649534225464, + "learning_rate": 1.0212253833336237e-05, + "loss": 0.7766, + "step": 537 + }, + { + "epoch": 1.6326220957799906, + "grad_norm": 2.217928409576416, + "learning_rate": 1.0176882253320968e-05, + "loss": 0.7529, + "step": 538 + }, + { + "epoch": 1.6356567093409198, + "grad_norm": 2.2367630004882812, + "learning_rate": 1.0141508459407622e-05, + "loss": 0.7699, + "step": 539 + }, + { + "epoch": 1.6386913229018494, + "grad_norm": 2.1163787841796875, + "learning_rate": 1.0106132894342564e-05, + "loss": 0.7637, + "step": 540 + }, + { + "epoch": 1.6417259364627785, + "grad_norm": 2.083292245864868, + "learning_rate": 1.0070756000894321e-05, + "loss": 0.783, + "step": 541 + }, + { + "epoch": 1.644760550023708, + "grad_norm": 2.1091578006744385, + "learning_rate": 1.0035378221848053e-05, + "loss": 0.7848, + "step": 542 + }, + { + "epoch": 1.6477951635846373, + "grad_norm": 2.0994367599487305, + "learning_rate": 1e-05, + "loss": 0.7698, + "step": 543 + }, + { + "epoch": 1.6508297771455667, + "grad_norm": 2.2539420127868652, + "learning_rate": 9.964621778151947e-06, + "loss": 0.769, + "step": 544 + }, + { + "epoch": 1.653864390706496, + "grad_norm": 2.154353380203247, + "learning_rate": 9.929243999105682e-06, + "loss": 0.7496, + "step": 545 + }, + { + "epoch": 1.6568990042674252, + "grad_norm": 2.265385150909424, + "learning_rate": 9.89386710565744e-06, + "loss": 0.768, + "step": 546 + }, + { + "epoch": 1.6599336178283548, + "grad_norm": 2.073289632797241, + "learning_rate": 9.858491540592383e-06, + "loss": 0.7807, + "step": 547 + }, + { + "epoch": 1.662968231389284, + "grad_norm": 2.135737657546997, + "learning_rate": 9.823117746679034e-06, + "loss": 0.766, + "step": 548 + }, + { + "epoch": 1.6660028449502133, + "grad_norm": 2.1423635482788086, + "learning_rate": 9.787746166663765e-06, + "loss": 0.7588, + "step": 549 + }, + { + "epoch": 1.6690374585111427, + "grad_norm": 2.0772769451141357, + "learning_rate": 9.752377243265229e-06, + "loss": 0.7647, + "step": 550 + }, + { + "epoch": 1.6690374585111427, + "eval_loss": NaN, + "eval_runtime": 204.4783, + "eval_samples_per_second": 20.335, + "eval_steps_per_second": 5.086, + "step": 550 + }, + { + "epoch": 1.672072072072072, + "grad_norm": 2.118854284286499, + "learning_rate": 9.71701141916882e-06, + "loss": 0.7728, + "step": 551 + }, + { + "epoch": 1.6751066856330015, + "grad_norm": 2.1113169193267822, + "learning_rate": 9.681649137021158e-06, + "loss": 0.7796, + "step": 552 + }, + { + "epoch": 1.6781412991939306, + "grad_norm": 2.064993143081665, + "learning_rate": 9.646290839424515e-06, + "loss": 0.762, + "step": 553 + }, + { + "epoch": 1.6811759127548602, + "grad_norm": 2.151028633117676, + "learning_rate": 9.610936968931302e-06, + "loss": 0.7709, + "step": 554 + }, + { + "epoch": 1.6842105263157894, + "grad_norm": 2.201429843902588, + "learning_rate": 9.57558796803852e-06, + "loss": 0.7619, + "step": 555 + }, + { + "epoch": 1.687245139876719, + "grad_norm": 2.065720796585083, + "learning_rate": 9.540244279182206e-06, + "loss": 0.798, + "step": 556 + }, + { + "epoch": 1.6902797534376481, + "grad_norm": 2.1895217895507812, + "learning_rate": 9.504906344731933e-06, + "loss": 0.7802, + "step": 557 + }, + { + "epoch": 1.6933143669985775, + "grad_norm": 2.2065324783325195, + "learning_rate": 9.46957460698523e-06, + "loss": 0.7767, + "step": 558 + }, + { + "epoch": 1.696348980559507, + "grad_norm": 2.17842698097229, + "learning_rate": 9.434249508162076e-06, + "loss": 0.7604, + "step": 559 + }, + { + "epoch": 1.6993835941204363, + "grad_norm": 2.164586305618286, + "learning_rate": 9.398931490399363e-06, + "loss": 0.756, + "step": 560 + }, + { + "epoch": 1.7024182076813656, + "grad_norm": 2.235797643661499, + "learning_rate": 9.363620995745337e-06, + "loss": 0.777, + "step": 561 + }, + { + "epoch": 1.7054528212422948, + "grad_norm": 2.084228038787842, + "learning_rate": 9.328318466154102e-06, + "loss": 0.765, + "step": 562 + }, + { + "epoch": 1.7084874348032244, + "grad_norm": 2.1382224559783936, + "learning_rate": 9.293024343480056e-06, + "loss": 0.7712, + "step": 563 + }, + { + "epoch": 1.7115220483641536, + "grad_norm": 2.304201126098633, + "learning_rate": 9.257739069472375e-06, + "loss": 0.7748, + "step": 564 + }, + { + "epoch": 1.714556661925083, + "grad_norm": 2.1342921257019043, + "learning_rate": 9.222463085769495e-06, + "loss": 0.762, + "step": 565 + }, + { + "epoch": 1.7175912754860123, + "grad_norm": 2.189845323562622, + "learning_rate": 9.187196833893559e-06, + "loss": 0.7605, + "step": 566 + }, + { + "epoch": 1.7206258890469417, + "grad_norm": 2.3078930377960205, + "learning_rate": 9.151940755244912e-06, + "loss": 0.7677, + "step": 567 + }, + { + "epoch": 1.723660502607871, + "grad_norm": 2.303213357925415, + "learning_rate": 9.11669529109656e-06, + "loss": 0.7735, + "step": 568 + }, + { + "epoch": 1.7266951161688002, + "grad_norm": 2.3082435131073, + "learning_rate": 9.081460882588668e-06, + "loss": 0.7683, + "step": 569 + }, + { + "epoch": 1.7297297297297298, + "grad_norm": 2.2995338439941406, + "learning_rate": 9.046237970723022e-06, + "loss": 0.752, + "step": 570 + }, + { + "epoch": 1.732764343290659, + "grad_norm": 2.0647215843200684, + "learning_rate": 9.011026996357504e-06, + "loss": 0.7731, + "step": 571 + }, + { + "epoch": 1.7357989568515886, + "grad_norm": 2.160496473312378, + "learning_rate": 8.975828400200592e-06, + "loss": 0.7543, + "step": 572 + }, + { + "epoch": 1.7388335704125177, + "grad_norm": 2.3624351024627686, + "learning_rate": 8.94064262280584e-06, + "loss": 0.774, + "step": 573 + }, + { + "epoch": 1.7418681839734471, + "grad_norm": 2.165922164916992, + "learning_rate": 8.90547010456635e-06, + "loss": 0.7815, + "step": 574 + }, + { + "epoch": 1.7449027975343765, + "grad_norm": 2.138921022415161, + "learning_rate": 8.870311285709274e-06, + "loss": 0.7593, + "step": 575 + }, + { + "epoch": 1.7479374110953059, + "grad_norm": 2.122746706008911, + "learning_rate": 8.835166606290295e-06, + "loss": 0.783, + "step": 576 + }, + { + "epoch": 1.7509720246562352, + "grad_norm": 2.119318723678589, + "learning_rate": 8.80003650618813e-06, + "loss": 0.7778, + "step": 577 + }, + { + "epoch": 1.7540066382171644, + "grad_norm": 2.092092990875244, + "learning_rate": 8.76492142509902e-06, + "loss": 0.765, + "step": 578 + }, + { + "epoch": 1.757041251778094, + "grad_norm": 2.156517267227173, + "learning_rate": 8.729821802531213e-06, + "loss": 0.7583, + "step": 579 + }, + { + "epoch": 1.7600758653390232, + "grad_norm": 2.2096939086914062, + "learning_rate": 8.694738077799487e-06, + "loss": 0.7698, + "step": 580 + }, + { + "epoch": 1.7631104788999525, + "grad_norm": 1.981307029724121, + "learning_rate": 8.659670690019626e-06, + "loss": 0.7771, + "step": 581 + }, + { + "epoch": 1.766145092460882, + "grad_norm": 2.1369874477386475, + "learning_rate": 8.624620078102952e-06, + "loss": 0.777, + "step": 582 + }, + { + "epoch": 1.7691797060218113, + "grad_norm": 2.1329455375671387, + "learning_rate": 8.58958668075081e-06, + "loss": 0.7748, + "step": 583 + }, + { + "epoch": 1.7722143195827407, + "grad_norm": 2.1630642414093018, + "learning_rate": 8.554570936449073e-06, + "loss": 0.7675, + "step": 584 + }, + { + "epoch": 1.7752489331436698, + "grad_norm": 2.1383938789367676, + "learning_rate": 8.519573283462688e-06, + "loss": 0.7702, + "step": 585 + }, + { + "epoch": 1.7782835467045994, + "grad_norm": 2.001569986343384, + "learning_rate": 8.484594159830142e-06, + "loss": 0.7553, + "step": 586 + }, + { + "epoch": 1.7813181602655286, + "grad_norm": 2.1266000270843506, + "learning_rate": 8.449634003358022e-06, + "loss": 0.7757, + "step": 587 + }, + { + "epoch": 1.7843527738264582, + "grad_norm": 2.172466278076172, + "learning_rate": 8.414693251615513e-06, + "loss": 0.7625, + "step": 588 + }, + { + "epoch": 1.7873873873873873, + "grad_norm": 2.2140755653381348, + "learning_rate": 8.379772341928916e-06, + "loss": 0.7653, + "step": 589 + }, + { + "epoch": 1.7904220009483167, + "grad_norm": 2.1517114639282227, + "learning_rate": 8.344871711376202e-06, + "loss": 0.7744, + "step": 590 + }, + { + "epoch": 1.793456614509246, + "grad_norm": 2.1749329566955566, + "learning_rate": 8.309991796781512e-06, + "loss": 0.7663, + "step": 591 + }, + { + "epoch": 1.7964912280701755, + "grad_norm": 2.2055773735046387, + "learning_rate": 8.2751330347097e-06, + "loss": 0.7768, + "step": 592 + }, + { + "epoch": 1.7995258416311048, + "grad_norm": 2.117532253265381, + "learning_rate": 8.24029586146088e-06, + "loss": 0.7774, + "step": 593 + }, + { + "epoch": 1.802560455192034, + "grad_norm": 2.2061290740966797, + "learning_rate": 8.205480713064947e-06, + "loss": 0.7819, + "step": 594 + }, + { + "epoch": 1.8055950687529636, + "grad_norm": 2.2168619632720947, + "learning_rate": 8.170688025276134e-06, + "loss": 0.7646, + "step": 595 + }, + { + "epoch": 1.8086296823138928, + "grad_norm": 2.238391637802124, + "learning_rate": 8.135918233567545e-06, + "loss": 0.7782, + "step": 596 + }, + { + "epoch": 1.8116642958748221, + "grad_norm": 2.105041265487671, + "learning_rate": 8.101171773125716e-06, + "loss": 0.7525, + "step": 597 + }, + { + "epoch": 1.8146989094357515, + "grad_norm": 2.0234577655792236, + "learning_rate": 8.066449078845168e-06, + "loss": 0.7658, + "step": 598 + }, + { + "epoch": 1.817733522996681, + "grad_norm": 2.188751220703125, + "learning_rate": 8.031750585322948e-06, + "loss": 0.7616, + "step": 599 + }, + { + "epoch": 1.8207681365576103, + "grad_norm": 2.1460800170898438, + "learning_rate": 7.99707672685321e-06, + "loss": 0.7617, + "step": 600 + }, + { + "epoch": 1.8207681365576103, + "eval_loss": NaN, + "eval_runtime": 204.6188, + "eval_samples_per_second": 20.321, + "eval_steps_per_second": 5.083, + "step": 600 + }, + { + "epoch": 1.8238027501185394, + "grad_norm": 1.964609146118164, + "learning_rate": 7.962427937421763e-06, + "loss": 0.7876, + "step": 601 + }, + { + "epoch": 1.826837363679469, + "grad_norm": 2.0918426513671875, + "learning_rate": 7.92780465070066e-06, + "loss": 0.7744, + "step": 602 + }, + { + "epoch": 1.8298719772403982, + "grad_norm": 2.059992790222168, + "learning_rate": 7.89320730004274e-06, + "loss": 0.7665, + "step": 603 + }, + { + "epoch": 1.8329065908013278, + "grad_norm": 2.0450618267059326, + "learning_rate": 7.858636318476226e-06, + "loss": 0.7636, + "step": 604 + }, + { + "epoch": 1.835941204362257, + "grad_norm": 2.1291024684906006, + "learning_rate": 7.824092138699307e-06, + "loss": 0.7652, + "step": 605 + }, + { + "epoch": 1.8389758179231863, + "grad_norm": 2.0631322860717773, + "learning_rate": 7.789575193074703e-06, + "loss": 0.7819, + "step": 606 + }, + { + "epoch": 1.8420104314841157, + "grad_norm": 2.1132242679595947, + "learning_rate": 7.755085913624274e-06, + "loss": 0.7507, + "step": 607 + }, + { + "epoch": 1.845045045045045, + "grad_norm": 2.1836376190185547, + "learning_rate": 7.720624732023604e-06, + "loss": 0.7798, + "step": 608 + }, + { + "epoch": 1.8480796586059745, + "grad_norm": 2.087453842163086, + "learning_rate": 7.686192079596586e-06, + "loss": 0.7817, + "step": 609 + }, + { + "epoch": 1.8511142721669036, + "grad_norm": 2.0749568939208984, + "learning_rate": 7.651788387310053e-06, + "loss": 0.7893, + "step": 610 + }, + { + "epoch": 1.8541488857278332, + "grad_norm": 2.085495710372925, + "learning_rate": 7.617414085768352e-06, + "loss": 0.7765, + "step": 611 + }, + { + "epoch": 1.8571834992887624, + "grad_norm": 2.1710901260375977, + "learning_rate": 7.5830696052079754e-06, + "loss": 0.7745, + "step": 612 + }, + { + "epoch": 1.860218112849692, + "grad_norm": 2.12780499458313, + "learning_rate": 7.548755375492173e-06, + "loss": 0.7845, + "step": 613 + }, + { + "epoch": 1.8632527264106211, + "grad_norm": 2.240459680557251, + "learning_rate": 7.514471826105556e-06, + "loss": 0.7731, + "step": 614 + }, + { + "epoch": 1.8662873399715505, + "grad_norm": 2.130253553390503, + "learning_rate": 7.480219386148751e-06, + "loss": 0.7904, + "step": 615 + }, + { + "epoch": 1.8693219535324799, + "grad_norm": 2.216806650161743, + "learning_rate": 7.445998484332994e-06, + "loss": 0.7529, + "step": 616 + }, + { + "epoch": 1.8723565670934093, + "grad_norm": 2.250821352005005, + "learning_rate": 7.411809548974792e-06, + "loss": 0.7566, + "step": 617 + }, + { + "epoch": 1.8753911806543386, + "grad_norm": 2.088843822479248, + "learning_rate": 7.377653007990559e-06, + "loss": 0.7568, + "step": 618 + }, + { + "epoch": 1.8784257942152678, + "grad_norm": 2.108356237411499, + "learning_rate": 7.343529288891239e-06, + "loss": 0.7769, + "step": 619 + }, + { + "epoch": 1.8814604077761974, + "grad_norm": 2.1201224327087402, + "learning_rate": 7.3094388187769815e-06, + "loss": 0.7651, + "step": 620 + }, + { + "epoch": 1.8844950213371265, + "grad_norm": 2.066652536392212, + "learning_rate": 7.275382024331773e-06, + "loss": 0.7773, + "step": 621 + }, + { + "epoch": 1.887529634898056, + "grad_norm": 2.1200613975524902, + "learning_rate": 7.241359331818121e-06, + "loss": 0.7835, + "step": 622 + }, + { + "epoch": 1.8905642484589853, + "grad_norm": 2.160304546356201, + "learning_rate": 7.2073711670717e-06, + "loss": 0.7489, + "step": 623 + }, + { + "epoch": 1.8935988620199147, + "grad_norm": 2.2403597831726074, + "learning_rate": 7.173417955496025e-06, + "loss": 0.7579, + "step": 624 + }, + { + "epoch": 1.896633475580844, + "grad_norm": 2.083029270172119, + "learning_rate": 7.13950012205713e-06, + "loss": 0.7715, + "step": 625 + }, + { + "epoch": 1.8996680891417732, + "grad_norm": 2.031341075897217, + "learning_rate": 7.105618091278246e-06, + "loss": 0.7494, + "step": 626 + }, + { + "epoch": 1.9027027027027028, + "grad_norm": 2.155595541000366, + "learning_rate": 7.071772287234497e-06, + "loss": 0.7797, + "step": 627 + }, + { + "epoch": 1.905737316263632, + "grad_norm": 2.2580630779266357, + "learning_rate": 7.037963133547583e-06, + "loss": 0.7801, + "step": 628 + }, + { + "epoch": 1.9087719298245616, + "grad_norm": 2.2416763305664062, + "learning_rate": 7.004191053380469e-06, + "loss": 0.7753, + "step": 629 + }, + { + "epoch": 1.9118065433854907, + "grad_norm": 2.0690417289733887, + "learning_rate": 6.970456469432116e-06, + "loss": 0.7693, + "step": 630 + }, + { + "epoch": 1.91484115694642, + "grad_norm": 2.0331270694732666, + "learning_rate": 6.936759803932167e-06, + "loss": 0.7578, + "step": 631 + }, + { + "epoch": 1.9178757705073495, + "grad_norm": 2.1094417572021484, + "learning_rate": 6.903101478635663e-06, + "loss": 0.7589, + "step": 632 + }, + { + "epoch": 1.9209103840682789, + "grad_norm": 2.042919635772705, + "learning_rate": 6.869481914817779e-06, + "loss": 0.7659, + "step": 633 + }, + { + "epoch": 1.9239449976292082, + "grad_norm": 2.0610294342041016, + "learning_rate": 6.835901533268536e-06, + "loss": 0.7648, + "step": 634 + }, + { + "epoch": 1.9269796111901374, + "grad_norm": 2.0473015308380127, + "learning_rate": 6.802360754287548e-06, + "loss": 0.7763, + "step": 635 + }, + { + "epoch": 1.930014224751067, + "grad_norm": 2.1137821674346924, + "learning_rate": 6.768859997678751e-06, + "loss": 0.7849, + "step": 636 + }, + { + "epoch": 1.9330488383119961, + "grad_norm": 2.12430477142334, + "learning_rate": 6.735399682745145e-06, + "loss": 0.7644, + "step": 637 + }, + { + "epoch": 1.9360834518729255, + "grad_norm": 2.0620853900909424, + "learning_rate": 6.701980228283569e-06, + "loss": 0.7733, + "step": 638 + }, + { + "epoch": 1.939118065433855, + "grad_norm": 2.124861478805542, + "learning_rate": 6.668602052579425e-06, + "loss": 0.7538, + "step": 639 + }, + { + "epoch": 1.9421526789947843, + "grad_norm": 2.003215789794922, + "learning_rate": 6.635265573401475e-06, + "loss": 0.7574, + "step": 640 + }, + { + "epoch": 1.9451872925557137, + "grad_norm": 2.1591827869415283, + "learning_rate": 6.601971207996592e-06, + "loss": 0.7591, + "step": 641 + }, + { + "epoch": 1.9482219061166428, + "grad_norm": 2.1029608249664307, + "learning_rate": 6.5687193730845375e-06, + "loss": 0.7927, + "step": 642 + }, + { + "epoch": 1.9512565196775724, + "grad_norm": 2.069796085357666, + "learning_rate": 6.535510484852767e-06, + "loss": 0.7491, + "step": 643 + }, + { + "epoch": 1.9542911332385016, + "grad_norm": 2.0876927375793457, + "learning_rate": 6.50234495895119e-06, + "loss": 0.7615, + "step": 644 + }, + { + "epoch": 1.9573257467994312, + "grad_norm": 2.0911192893981934, + "learning_rate": 6.469223210486992e-06, + "loss": 0.7579, + "step": 645 + }, + { + "epoch": 1.9603603603603603, + "grad_norm": 2.2714855670928955, + "learning_rate": 6.4361456540194325e-06, + "loss": 0.7543, + "step": 646 + }, + { + "epoch": 1.9633949739212897, + "grad_norm": 2.075011968612671, + "learning_rate": 6.403112703554643e-06, + "loss": 0.752, + "step": 647 + }, + { + "epoch": 1.966429587482219, + "grad_norm": 2.097029447555542, + "learning_rate": 6.370124772540469e-06, + "loss": 0.7338, + "step": 648 + }, + { + "epoch": 1.9694642010431485, + "grad_norm": 2.147951602935791, + "learning_rate": 6.337182273861273e-06, + "loss": 0.7735, + "step": 649 + }, + { + "epoch": 1.9724988146040778, + "grad_norm": 2.0643298625946045, + "learning_rate": 6.3042856198327795e-06, + "loss": 0.7775, + "step": 650 + }, + { + "epoch": 1.9724988146040778, + "eval_loss": NaN, + "eval_runtime": 204.7127, + "eval_samples_per_second": 20.311, + "eval_steps_per_second": 5.08, + "step": 650 + }, + { + "epoch": 1.975533428165007, + "grad_norm": 2.043440580368042, + "learning_rate": 6.2714352221969155e-06, + "loss": 0.7593, + "step": 651 + }, + { + "epoch": 1.9785680417259366, + "grad_norm": 2.250242233276367, + "learning_rate": 6.238631492116643e-06, + "loss": 0.7657, + "step": 652 + }, + { + "epoch": 1.9816026552868657, + "grad_norm": 2.068997383117676, + "learning_rate": 6.205874840170833e-06, + "loss": 0.7774, + "step": 653 + }, + { + "epoch": 1.9846372688477951, + "grad_norm": 2.1658010482788086, + "learning_rate": 6.173165676349103e-06, + "loss": 0.7792, + "step": 654 + }, + { + "epoch": 1.9876718824087245, + "grad_norm": 2.1947412490844727, + "learning_rate": 6.140504410046712e-06, + "loss": 0.7631, + "step": 655 + }, + { + "epoch": 1.9907064959696539, + "grad_norm": 2.082087278366089, + "learning_rate": 6.107891450059419e-06, + "loss": 0.7564, + "step": 656 + }, + { + "epoch": 1.9937411095305833, + "grad_norm": 2.043905019760132, + "learning_rate": 6.075327204578363e-06, + "loss": 0.7834, + "step": 657 + }, + { + "epoch": 1.9967757230915124, + "grad_norm": 2.1089298725128174, + "learning_rate": 6.042812081184973e-06, + "loss": 0.7595, + "step": 658 + }, + { + "epoch": 1.999810336652442, + "grad_norm": 2.0419466495513916, + "learning_rate": 6.010346486845837e-06, + "loss": 0.7481, + "step": 659 + }, + { + "epoch": 2.002844950213371, + "grad_norm": 2.1154048442840576, + "learning_rate": 5.97793082790765e-06, + "loss": 0.7479, + "step": 660 + }, + { + "epoch": 2.0058795637743008, + "grad_norm": 1.9878915548324585, + "learning_rate": 5.945565510092086e-06, + "loss": 0.7547, + "step": 661 + }, + { + "epoch": 2.00891417733523, + "grad_norm": 2.1136927604675293, + "learning_rate": 5.913250938490744e-06, + "loss": 0.7294, + "step": 662 + }, + { + "epoch": 2.0119487908961595, + "grad_norm": 2.1599817276000977, + "learning_rate": 5.880987517560075e-06, + "loss": 0.744, + "step": 663 + }, + { + "epoch": 2.0149834044570887, + "grad_norm": 2.200779438018799, + "learning_rate": 5.84877565111631e-06, + "loss": 0.7353, + "step": 664 + }, + { + "epoch": 2.018018018018018, + "grad_norm": 2.05635404586792, + "learning_rate": 5.81661574233042e-06, + "loss": 0.7534, + "step": 665 + }, + { + "epoch": 2.0210526315789474, + "grad_norm": 2.193406820297241, + "learning_rate": 5.784508193723058e-06, + "loss": 0.7443, + "step": 666 + }, + { + "epoch": 2.0240872451398766, + "grad_norm": 2.120541572570801, + "learning_rate": 5.752453407159521e-06, + "loss": 0.7414, + "step": 667 + }, + { + "epoch": 2.027121858700806, + "grad_norm": 2.075017213821411, + "learning_rate": 5.720451783844741e-06, + "loss": 0.7439, + "step": 668 + }, + { + "epoch": 2.0301564722617353, + "grad_norm": 2.1643104553222656, + "learning_rate": 5.688503724318217e-06, + "loss": 0.7565, + "step": 669 + }, + { + "epoch": 2.033191085822665, + "grad_norm": 2.148364782333374, + "learning_rate": 5.656609628449064e-06, + "loss": 0.7353, + "step": 670 + }, + { + "epoch": 2.036225699383594, + "grad_norm": 2.205618143081665, + "learning_rate": 5.6247698954309616e-06, + "loss": 0.7529, + "step": 671 + }, + { + "epoch": 2.0392603129445233, + "grad_norm": 2.1727452278137207, + "learning_rate": 5.592984923777156e-06, + "loss": 0.7439, + "step": 672 + }, + { + "epoch": 2.042294926505453, + "grad_norm": 2.1897435188293457, + "learning_rate": 5.561255111315525e-06, + "loss": 0.7578, + "step": 673 + }, + { + "epoch": 2.045329540066382, + "grad_norm": 2.155768871307373, + "learning_rate": 5.529580855183518e-06, + "loss": 0.7687, + "step": 674 + }, + { + "epoch": 2.0483641536273116, + "grad_norm": 2.100489616394043, + "learning_rate": 5.497962551823266e-06, + "loss": 0.7333, + "step": 675 + }, + { + "epoch": 2.0513987671882408, + "grad_norm": 2.0687026977539062, + "learning_rate": 5.466400596976568e-06, + "loss": 0.745, + "step": 676 + }, + { + "epoch": 2.0544333807491704, + "grad_norm": 2.144482135772705, + "learning_rate": 5.434895385679937e-06, + "loss": 0.7369, + "step": 677 + }, + { + "epoch": 2.0574679943100995, + "grad_norm": 2.1690595149993896, + "learning_rate": 5.403447312259702e-06, + "loss": 0.7443, + "step": 678 + }, + { + "epoch": 2.060502607871029, + "grad_norm": 2.0905041694641113, + "learning_rate": 5.3720567703270135e-06, + "loss": 0.7333, + "step": 679 + }, + { + "epoch": 2.0635372214319583, + "grad_norm": 2.205411911010742, + "learning_rate": 5.340724152772956e-06, + "loss": 0.7399, + "step": 680 + }, + { + "epoch": 2.0665718349928874, + "grad_norm": 2.3007540702819824, + "learning_rate": 5.3094498517636324e-06, + "loss": 0.7452, + "step": 681 + }, + { + "epoch": 2.069606448553817, + "grad_norm": 2.2882180213928223, + "learning_rate": 5.278234258735215e-06, + "loss": 0.7285, + "step": 682 + }, + { + "epoch": 2.072641062114746, + "grad_norm": 2.2169337272644043, + "learning_rate": 5.247077764389099e-06, + "loss": 0.7437, + "step": 683 + }, + { + "epoch": 2.075675675675676, + "grad_norm": 2.1462368965148926, + "learning_rate": 5.215980758686978e-06, + "loss": 0.7366, + "step": 684 + }, + { + "epoch": 2.078710289236605, + "grad_norm": 2.200030565261841, + "learning_rate": 5.18494363084596e-06, + "loss": 0.7546, + "step": 685 + }, + { + "epoch": 2.0817449027975345, + "grad_norm": 2.1369688510894775, + "learning_rate": 5.153966769333734e-06, + "loss": 0.7388, + "step": 686 + }, + { + "epoch": 2.0847795163584637, + "grad_norm": 2.191958427429199, + "learning_rate": 5.1230505618636575e-06, + "loss": 0.7297, + "step": 687 + }, + { + "epoch": 2.0878141299193933, + "grad_norm": 2.1999082565307617, + "learning_rate": 5.092195395389937e-06, + "loss": 0.7428, + "step": 688 + }, + { + "epoch": 2.0908487434803225, + "grad_norm": 2.373140335083008, + "learning_rate": 5.061401656102791e-06, + "loss": 0.7585, + "step": 689 + }, + { + "epoch": 2.0938833570412516, + "grad_norm": 2.2395219802856445, + "learning_rate": 5.030669729423572e-06, + "loss": 0.7382, + "step": 690 + }, + { + "epoch": 2.096917970602181, + "grad_norm": 2.1638119220733643, + "learning_rate": 5.000000000000003e-06, + "loss": 0.7538, + "step": 691 + }, + { + "epoch": 2.0999525841631104, + "grad_norm": 2.173945188522339, + "learning_rate": 4.969392851701306e-06, + "loss": 0.7403, + "step": 692 + }, + { + "epoch": 2.10298719772404, + "grad_norm": 2.2739264965057373, + "learning_rate": 4.938848667613436e-06, + "loss": 0.7459, + "step": 693 + }, + { + "epoch": 2.106021811284969, + "grad_norm": 2.2303757667541504, + "learning_rate": 4.908367830034284e-06, + "loss": 0.717, + "step": 694 + }, + { + "epoch": 2.1090564248458987, + "grad_norm": 2.083385944366455, + "learning_rate": 4.8779507204688595e-06, + "loss": 0.747, + "step": 695 + }, + { + "epoch": 2.112091038406828, + "grad_norm": 2.252095937728882, + "learning_rate": 4.84759771962455e-06, + "loss": 0.7414, + "step": 696 + }, + { + "epoch": 2.115125651967757, + "grad_norm": 2.192476272583008, + "learning_rate": 4.817309207406347e-06, + "loss": 0.7078, + "step": 697 + }, + { + "epoch": 2.1181602655286866, + "grad_norm": 2.170762062072754, + "learning_rate": 4.787085562912076e-06, + "loss": 0.7482, + "step": 698 + }, + { + "epoch": 2.121194879089616, + "grad_norm": 2.1275007724761963, + "learning_rate": 4.756927164427685e-06, + "loss": 0.7415, + "step": 699 + }, + { + "epoch": 2.1242294926505454, + "grad_norm": 2.1509687900543213, + "learning_rate": 4.726834389422461e-06, + "loss": 0.7326, + "step": 700 + }, + { + "epoch": 2.1242294926505454, + "eval_loss": NaN, + "eval_runtime": 204.0742, + "eval_samples_per_second": 20.375, + "eval_steps_per_second": 5.096, + "step": 700 + }, + { + "epoch": 2.1272641062114745, + "grad_norm": 2.0226032733917236, + "learning_rate": 4.696807614544352e-06, + "loss": 0.7411, + "step": 701 + }, + { + "epoch": 2.130298719772404, + "grad_norm": 2.1866512298583984, + "learning_rate": 4.666847215615225e-06, + "loss": 0.7394, + "step": 702 + }, + { + "epoch": 2.1333333333333333, + "grad_norm": 2.2013630867004395, + "learning_rate": 4.636953567626176e-06, + "loss": 0.7356, + "step": 703 + }, + { + "epoch": 2.1363679468942625, + "grad_norm": 2.4076430797576904, + "learning_rate": 4.607127044732827e-06, + "loss": 0.7501, + "step": 704 + }, + { + "epoch": 2.139402560455192, + "grad_norm": 2.1296753883361816, + "learning_rate": 4.57736802025065e-06, + "loss": 0.7399, + "step": 705 + }, + { + "epoch": 2.142437174016121, + "grad_norm": 2.22512149810791, + "learning_rate": 4.5476768666502895e-06, + "loss": 0.7553, + "step": 706 + }, + { + "epoch": 2.145471787577051, + "grad_norm": 2.178199052810669, + "learning_rate": 4.518053955552903e-06, + "loss": 0.7458, + "step": 707 + }, + { + "epoch": 2.14850640113798, + "grad_norm": 2.0975606441497803, + "learning_rate": 4.488499657725511e-06, + "loss": 0.7234, + "step": 708 + }, + { + "epoch": 2.1515410146989096, + "grad_norm": 2.1103413105010986, + "learning_rate": 4.459014343076356e-06, + "loss": 0.7431, + "step": 709 + }, + { + "epoch": 2.1545756282598387, + "grad_norm": 2.2594456672668457, + "learning_rate": 4.429598380650266e-06, + "loss": 0.7578, + "step": 710 + }, + { + "epoch": 2.1576102418207683, + "grad_norm": 2.2270963191986084, + "learning_rate": 4.400252138624047e-06, + "loss": 0.7461, + "step": 711 + }, + { + "epoch": 2.1606448553816975, + "grad_norm": 2.191631555557251, + "learning_rate": 4.370975984301866e-06, + "loss": 0.7425, + "step": 712 + }, + { + "epoch": 2.1636794689426266, + "grad_norm": 2.124616861343384, + "learning_rate": 4.341770284110655e-06, + "loss": 0.7495, + "step": 713 + }, + { + "epoch": 2.1667140825035562, + "grad_norm": 2.18497371673584, + "learning_rate": 4.312635403595532e-06, + "loss": 0.7239, + "step": 714 + }, + { + "epoch": 2.1697486960644854, + "grad_norm": 2.149658441543579, + "learning_rate": 4.283571707415214e-06, + "loss": 0.7325, + "step": 715 + }, + { + "epoch": 2.172783309625415, + "grad_norm": 2.133171558380127, + "learning_rate": 4.25457955933746e-06, + "loss": 0.7324, + "step": 716 + }, + { + "epoch": 2.175817923186344, + "grad_norm": 2.15606689453125, + "learning_rate": 4.2256593222345185e-06, + "loss": 0.7273, + "step": 717 + }, + { + "epoch": 2.1788525367472737, + "grad_norm": 2.237285614013672, + "learning_rate": 4.196811358078585e-06, + "loss": 0.7463, + "step": 718 + }, + { + "epoch": 2.181887150308203, + "grad_norm": 2.267974853515625, + "learning_rate": 4.168036027937267e-06, + "loss": 0.7405, + "step": 719 + }, + { + "epoch": 2.1849217638691325, + "grad_norm": 2.0772793292999268, + "learning_rate": 4.139333691969071e-06, + "loss": 0.7418, + "step": 720 + }, + { + "epoch": 2.1879563774300617, + "grad_norm": 2.1171600818634033, + "learning_rate": 4.1107047094188946e-06, + "loss": 0.7222, + "step": 721 + }, + { + "epoch": 2.190990990990991, + "grad_norm": 2.3039426803588867, + "learning_rate": 4.082149438613514e-06, + "loss": 0.731, + "step": 722 + }, + { + "epoch": 2.1940256045519204, + "grad_norm": 2.207404375076294, + "learning_rate": 4.053668236957135e-06, + "loss": 0.7499, + "step": 723 + }, + { + "epoch": 2.1970602181128496, + "grad_norm": 2.1732053756713867, + "learning_rate": 4.025261460926877e-06, + "loss": 0.7617, + "step": 724 + }, + { + "epoch": 2.200094831673779, + "grad_norm": 2.537623167037964, + "learning_rate": 3.996929466068344e-06, + "loss": 0.72, + "step": 725 + }, + { + "epoch": 2.2031294452347083, + "grad_norm": 2.2418384552001953, + "learning_rate": 3.96867260699116e-06, + "loss": 0.727, + "step": 726 + }, + { + "epoch": 2.206164058795638, + "grad_norm": 2.189655065536499, + "learning_rate": 3.940491237364519e-06, + "loss": 0.7321, + "step": 727 + }, + { + "epoch": 2.209198672356567, + "grad_norm": 2.1620960235595703, + "learning_rate": 3.912385709912794e-06, + "loss": 0.7421, + "step": 728 + }, + { + "epoch": 2.2122332859174962, + "grad_norm": 2.216374397277832, + "learning_rate": 3.884356376411089e-06, + "loss": 0.7632, + "step": 729 + }, + { + "epoch": 2.215267899478426, + "grad_norm": 2.152038335800171, + "learning_rate": 3.8564035876808405e-06, + "loss": 0.7349, + "step": 730 + }, + { + "epoch": 2.218302513039355, + "grad_norm": 2.1349170207977295, + "learning_rate": 3.828527693585451e-06, + "loss": 0.7373, + "step": 731 + }, + { + "epoch": 2.2213371266002846, + "grad_norm": 2.1102511882781982, + "learning_rate": 3.8007290430258712e-06, + "loss": 0.7278, + "step": 732 + }, + { + "epoch": 2.2243717401612138, + "grad_norm": 2.124293088912964, + "learning_rate": 3.7730079839362755e-06, + "loss": 0.7315, + "step": 733 + }, + { + "epoch": 2.2274063537221434, + "grad_norm": 2.0747625827789307, + "learning_rate": 3.7453648632796745e-06, + "loss": 0.7391, + "step": 734 + }, + { + "epoch": 2.2304409672830725, + "grad_norm": 2.1584184169769287, + "learning_rate": 3.7178000270435765e-06, + "loss": 0.7526, + "step": 735 + }, + { + "epoch": 2.2334755808440017, + "grad_norm": 2.0907371044158936, + "learning_rate": 3.690313820235686e-06, + "loss": 0.7176, + "step": 736 + }, + { + "epoch": 2.2365101944049313, + "grad_norm": 2.100431442260742, + "learning_rate": 3.662906586879542e-06, + "loss": 0.7619, + "step": 737 + }, + { + "epoch": 2.2395448079658604, + "grad_norm": 2.1690993309020996, + "learning_rate": 3.6355786700102426e-06, + "loss": 0.7385, + "step": 738 + }, + { + "epoch": 2.24257942152679, + "grad_norm": 2.2165980339050293, + "learning_rate": 3.6083304116701535e-06, + "loss": 0.7577, + "step": 739 + }, + { + "epoch": 2.245614035087719, + "grad_norm": 2.1825928688049316, + "learning_rate": 3.5811621529045927e-06, + "loss": 0.7353, + "step": 740 + }, + { + "epoch": 2.2486486486486488, + "grad_norm": 2.1485071182250977, + "learning_rate": 3.5540742337576083e-06, + "loss": 0.7534, + "step": 741 + }, + { + "epoch": 2.251683262209578, + "grad_norm": 2.193574905395508, + "learning_rate": 3.5270669932676926e-06, + "loss": 0.7395, + "step": 742 + }, + { + "epoch": 2.2547178757705075, + "grad_norm": 2.152099847793579, + "learning_rate": 3.5001407694635326e-06, + "loss": 0.7265, + "step": 743 + }, + { + "epoch": 2.2577524893314367, + "grad_norm": 2.2395544052124023, + "learning_rate": 3.4732958993598153e-06, + "loss": 0.7482, + "step": 744 + }, + { + "epoch": 2.2607871028923663, + "grad_norm": 2.209214448928833, + "learning_rate": 3.446532718952966e-06, + "loss": 0.7366, + "step": 745 + }, + { + "epoch": 2.2638217164532954, + "grad_norm": 2.282007932662964, + "learning_rate": 3.4198515632169703e-06, + "loss": 0.7641, + "step": 746 + }, + { + "epoch": 2.2668563300142246, + "grad_norm": 2.1614651679992676, + "learning_rate": 3.3932527660991877e-06, + "loss": 0.7502, + "step": 747 + }, + { + "epoch": 2.269890943575154, + "grad_norm": 2.306121826171875, + "learning_rate": 3.3667366605161322e-06, + "loss": 0.7464, + "step": 748 + }, + { + "epoch": 2.2729255571360834, + "grad_norm": 2.2220146656036377, + "learning_rate": 3.340303578349361e-06, + "loss": 0.7497, + "step": 749 + }, + { + "epoch": 2.275960170697013, + "grad_norm": 2.1184372901916504, + "learning_rate": 3.313953850441266e-06, + "loss": 0.7434, + "step": 750 + }, + { + "epoch": 2.275960170697013, + "eval_loss": NaN, + "eval_runtime": 203.9554, + "eval_samples_per_second": 20.387, + "eval_steps_per_second": 5.099, + "step": 750 + }, + { + "epoch": 2.278994784257942, + "grad_norm": 2.2895419597625732, + "learning_rate": 3.2876878065909714e-06, + "loss": 0.7362, + "step": 751 + }, + { + "epoch": 2.2820293978188717, + "grad_norm": 2.148529291152954, + "learning_rate": 3.2615057755502e-06, + "loss": 0.7558, + "step": 752 + }, + { + "epoch": 2.285064011379801, + "grad_norm": 2.1223018169403076, + "learning_rate": 3.2354080850191328e-06, + "loss": 0.7591, + "step": 753 + }, + { + "epoch": 2.28809862494073, + "grad_norm": 2.2450056076049805, + "learning_rate": 3.2093950616423397e-06, + "loss": 0.7335, + "step": 754 + }, + { + "epoch": 2.2911332385016596, + "grad_norm": 2.173351526260376, + "learning_rate": 3.1834670310046735e-06, + "loss": 0.7546, + "step": 755 + }, + { + "epoch": 2.2941678520625888, + "grad_norm": 2.0341904163360596, + "learning_rate": 3.157624317627195e-06, + "loss": 0.7447, + "step": 756 + }, + { + "epoch": 2.2972024656235184, + "grad_norm": 2.0900380611419678, + "learning_rate": 3.1318672449631283e-06, + "loss": 0.7544, + "step": 757 + }, + { + "epoch": 2.3002370791844475, + "grad_norm": 2.2536418437957764, + "learning_rate": 3.106196135393782e-06, + "loss": 0.7326, + "step": 758 + }, + { + "epoch": 2.303271692745377, + "grad_norm": 2.1906511783599854, + "learning_rate": 3.0806113102245395e-06, + "loss": 0.7522, + "step": 759 + }, + { + "epoch": 2.3063063063063063, + "grad_norm": 2.1476638317108154, + "learning_rate": 3.055113089680829e-06, + "loss": 0.7423, + "step": 760 + }, + { + "epoch": 2.3093409198672354, + "grad_norm": 2.1808698177337646, + "learning_rate": 3.029701792904117e-06, + "loss": 0.7612, + "step": 761 + }, + { + "epoch": 2.312375533428165, + "grad_norm": 2.085947036743164, + "learning_rate": 3.00437773794791e-06, + "loss": 0.7439, + "step": 762 + }, + { + "epoch": 2.315410146989094, + "grad_norm": 2.1122984886169434, + "learning_rate": 2.979141241773775e-06, + "loss": 0.7383, + "step": 763 + }, + { + "epoch": 2.318444760550024, + "grad_norm": 2.1743266582489014, + "learning_rate": 2.953992620247379e-06, + "loss": 0.7458, + "step": 764 + }, + { + "epoch": 2.321479374110953, + "grad_norm": 2.1469690799713135, + "learning_rate": 2.9289321881345257e-06, + "loss": 0.756, + "step": 765 + }, + { + "epoch": 2.3245139876718826, + "grad_norm": 2.077439069747925, + "learning_rate": 2.9039602590972203e-06, + "loss": 0.7316, + "step": 766 + }, + { + "epoch": 2.3275486012328117, + "grad_norm": 2.4407780170440674, + "learning_rate": 2.879077145689746e-06, + "loss": 0.7343, + "step": 767 + }, + { + "epoch": 2.330583214793741, + "grad_norm": 2.1479241847991943, + "learning_rate": 2.8542831593547483e-06, + "loss": 0.7407, + "step": 768 + }, + { + "epoch": 2.3336178283546705, + "grad_norm": 2.2073941230773926, + "learning_rate": 2.829578610419337e-06, + "loss": 0.7471, + "step": 769 + }, + { + "epoch": 2.3366524419155996, + "grad_norm": 2.0292317867279053, + "learning_rate": 2.8049638080912045e-06, + "loss": 0.7332, + "step": 770 + }, + { + "epoch": 2.339687055476529, + "grad_norm": 2.2681589126586914, + "learning_rate": 2.780439060454756e-06, + "loss": 0.7436, + "step": 771 + }, + { + "epoch": 2.3427216690374584, + "grad_norm": 2.057676076889038, + "learning_rate": 2.75600467446725e-06, + "loss": 0.7352, + "step": 772 + }, + { + "epoch": 2.345756282598388, + "grad_norm": 2.2097206115722656, + "learning_rate": 2.7316609559549568e-06, + "loss": 0.7275, + "step": 773 + }, + { + "epoch": 2.348790896159317, + "grad_norm": 2.3039231300354004, + "learning_rate": 2.707408209609339e-06, + "loss": 0.7556, + "step": 774 + }, + { + "epoch": 2.3518255097202467, + "grad_norm": 2.097167730331421, + "learning_rate": 2.683246738983217e-06, + "loss": 0.7404, + "step": 775 + }, + { + "epoch": 2.354860123281176, + "grad_norm": 2.083677053451538, + "learning_rate": 2.6591768464870016e-06, + "loss": 0.7359, + "step": 776 + }, + { + "epoch": 2.3578947368421055, + "grad_norm": 2.0094263553619385, + "learning_rate": 2.6351988333848787e-06, + "loss": 0.7297, + "step": 777 + }, + { + "epoch": 2.3609293504030346, + "grad_norm": 2.13173770904541, + "learning_rate": 2.611312999791055e-06, + "loss": 0.7315, + "step": 778 + }, + { + "epoch": 2.363963963963964, + "grad_norm": 2.1343095302581787, + "learning_rate": 2.587519644666001e-06, + "loss": 0.7309, + "step": 779 + }, + { + "epoch": 2.3669985775248934, + "grad_norm": 2.122774600982666, + "learning_rate": 2.5638190658126937e-06, + "loss": 0.7428, + "step": 780 + }, + { + "epoch": 2.3700331910858226, + "grad_norm": 2.1483750343322754, + "learning_rate": 2.5402115598729182e-06, + "loss": 0.7386, + "step": 781 + }, + { + "epoch": 2.373067804646752, + "grad_norm": 2.189150810241699, + "learning_rate": 2.5166974223235295e-06, + "loss": 0.7237, + "step": 782 + }, + { + "epoch": 2.3761024182076813, + "grad_norm": 2.156003475189209, + "learning_rate": 2.493276947472756e-06, + "loss": 0.7614, + "step": 783 + }, + { + "epoch": 2.379137031768611, + "grad_norm": 2.100486993789673, + "learning_rate": 2.4699504284565413e-06, + "loss": 0.7543, + "step": 784 + }, + { + "epoch": 2.38217164532954, + "grad_norm": 2.1420738697052, + "learning_rate": 2.446718157234832e-06, + "loss": 0.7475, + "step": 785 + }, + { + "epoch": 2.385206258890469, + "grad_norm": 2.205798625946045, + "learning_rate": 2.4235804245879723e-06, + "loss": 0.7362, + "step": 786 + }, + { + "epoch": 2.388240872451399, + "grad_norm": 2.193894147872925, + "learning_rate": 2.4005375201130275e-06, + "loss": 0.7456, + "step": 787 + }, + { + "epoch": 2.391275486012328, + "grad_norm": 2.1630735397338867, + "learning_rate": 2.3775897322201667e-06, + "loss": 0.7553, + "step": 788 + }, + { + "epoch": 2.3943100995732576, + "grad_norm": 2.247066020965576, + "learning_rate": 2.354737348129077e-06, + "loss": 0.7324, + "step": 789 + }, + { + "epoch": 2.3973447131341867, + "grad_norm": 2.1332616806030273, + "learning_rate": 2.33198065386533e-06, + "loss": 0.7413, + "step": 790 + }, + { + "epoch": 2.4003793266951163, + "grad_norm": 2.1438822746276855, + "learning_rate": 2.3093199342568316e-06, + "loss": 0.7215, + "step": 791 + }, + { + "epoch": 2.4034139402560455, + "grad_norm": 2.1765286922454834, + "learning_rate": 2.2867554729302545e-06, + "loss": 0.7362, + "step": 792 + }, + { + "epoch": 2.4064485538169746, + "grad_norm": 2.059136390686035, + "learning_rate": 2.2642875523074613e-06, + "loss": 0.7373, + "step": 793 + }, + { + "epoch": 2.4094831673779042, + "grad_norm": 2.195065975189209, + "learning_rate": 2.2419164536020112e-06, + "loss": 0.7363, + "step": 794 + }, + { + "epoch": 2.4125177809388334, + "grad_norm": 2.2990760803222656, + "learning_rate": 2.2196424568156073e-06, + "loss": 0.7348, + "step": 795 + }, + { + "epoch": 2.415552394499763, + "grad_norm": 2.251936435699463, + "learning_rate": 2.197465840734596e-06, + "loss": 0.7257, + "step": 796 + }, + { + "epoch": 2.418587008060692, + "grad_norm": 2.142132520675659, + "learning_rate": 2.1753868829265046e-06, + "loss": 0.7526, + "step": 797 + }, + { + "epoch": 2.4216216216216218, + "grad_norm": 2.235971450805664, + "learning_rate": 2.1534058597365284e-06, + "loss": 0.7402, + "step": 798 + }, + { + "epoch": 2.424656235182551, + "grad_norm": 2.1040003299713135, + "learning_rate": 2.1315230462840985e-06, + "loss": 0.7271, + "step": 799 + }, + { + "epoch": 2.4276908487434805, + "grad_norm": 2.153122663497925, + "learning_rate": 2.1097387164594406e-06, + "loss": 0.7274, + "step": 800 + }, + { + "epoch": 2.4276908487434805, + "eval_loss": NaN, + "eval_runtime": 204.5921, + "eval_samples_per_second": 20.323, + "eval_steps_per_second": 5.083, + "step": 800 + }, + { + "epoch": 2.4307254623044097, + "grad_norm": 2.2538816928863525, + "learning_rate": 2.0880531429201146e-06, + "loss": 0.749, + "step": 801 + }, + { + "epoch": 2.4337600758653393, + "grad_norm": 2.194967031478882, + "learning_rate": 2.0664665970876496e-06, + "loss": 0.7395, + "step": 802 + }, + { + "epoch": 2.4367946894262684, + "grad_norm": 2.1779520511627197, + "learning_rate": 2.0449793491441026e-06, + "loss": 0.7476, + "step": 803 + }, + { + "epoch": 2.4398293029871976, + "grad_norm": 2.243781805038452, + "learning_rate": 2.0235916680287015e-06, + "loss": 0.7268, + "step": 804 + }, + { + "epoch": 2.442863916548127, + "grad_norm": 2.1497793197631836, + "learning_rate": 2.0023038214344827e-06, + "loss": 0.76, + "step": 805 + }, + { + "epoch": 2.4458985301090563, + "grad_norm": 2.209360122680664, + "learning_rate": 1.9811160758049163e-06, + "loss": 0.7133, + "step": 806 + }, + { + "epoch": 2.448933143669986, + "grad_norm": 2.1285481452941895, + "learning_rate": 1.960028696330596e-06, + "loss": 0.7349, + "step": 807 + }, + { + "epoch": 2.451967757230915, + "grad_norm": 2.16249680519104, + "learning_rate": 1.9390419469459066e-06, + "loss": 0.7392, + "step": 808 + }, + { + "epoch": 2.4550023707918447, + "grad_norm": 2.1789779663085938, + "learning_rate": 1.9181560903257234e-06, + "loss": 0.7299, + "step": 809 + }, + { + "epoch": 2.458036984352774, + "grad_norm": 2.196904420852661, + "learning_rate": 1.8973713878821343e-06, + "loss": 0.74, + "step": 810 + }, + { + "epoch": 2.461071597913703, + "grad_norm": 2.087092399597168, + "learning_rate": 1.8766880997611424e-06, + "loss": 0.7503, + "step": 811 + }, + { + "epoch": 2.4641062114746326, + "grad_norm": 2.023298978805542, + "learning_rate": 1.8561064848394384e-06, + "loss": 0.7409, + "step": 812 + }, + { + "epoch": 2.4671408250355618, + "grad_norm": 2.2470693588256836, + "learning_rate": 1.8356268007211442e-06, + "loss": 0.7553, + "step": 813 + }, + { + "epoch": 2.4701754385964914, + "grad_norm": 2.0922911167144775, + "learning_rate": 1.8152493037345942e-06, + "loss": 0.7257, + "step": 814 + }, + { + "epoch": 2.4732100521574205, + "grad_norm": 2.100867986679077, + "learning_rate": 1.7949742489291256e-06, + "loss": 0.7275, + "step": 815 + }, + { + "epoch": 2.47624466571835, + "grad_norm": 2.309476375579834, + "learning_rate": 1.7748018900718856e-06, + "loss": 0.7349, + "step": 816 + }, + { + "epoch": 2.4792792792792793, + "grad_norm": 2.1686556339263916, + "learning_rate": 1.7547324796446553e-06, + "loss": 0.746, + "step": 817 + }, + { + "epoch": 2.4823138928402084, + "grad_norm": 2.161126136779785, + "learning_rate": 1.7347662688406907e-06, + "loss": 0.7637, + "step": 818 + }, + { + "epoch": 2.485348506401138, + "grad_norm": 2.0931811332702637, + "learning_rate": 1.7149035075615795e-06, + "loss": 0.7417, + "step": 819 + }, + { + "epoch": 2.488383119962067, + "grad_norm": 2.220621347427368, + "learning_rate": 1.6951444444141084e-06, + "loss": 0.7426, + "step": 820 + }, + { + "epoch": 2.491417733522997, + "grad_norm": 2.1453609466552734, + "learning_rate": 1.6754893267071593e-06, + "loss": 0.7277, + "step": 821 + }, + { + "epoch": 2.494452347083926, + "grad_norm": 2.204213857650757, + "learning_rate": 1.6559384004486057e-06, + "loss": 0.7456, + "step": 822 + }, + { + "epoch": 2.4974869606448555, + "grad_norm": 2.1252691745758057, + "learning_rate": 1.6364919103422394e-06, + "loss": 0.7423, + "step": 823 + }, + { + "epoch": 2.5005215742057847, + "grad_norm": 2.285757064819336, + "learning_rate": 1.6171500997847056e-06, + "loss": 0.7377, + "step": 824 + }, + { + "epoch": 2.503556187766714, + "grad_norm": 2.1519298553466797, + "learning_rate": 1.5979132108624572e-06, + "loss": 0.7397, + "step": 825 + }, + { + "epoch": 2.5065908013276434, + "grad_norm": 2.1684153079986572, + "learning_rate": 1.5787814843487226e-06, + "loss": 0.7355, + "step": 826 + }, + { + "epoch": 2.509625414888573, + "grad_norm": 2.1787893772125244, + "learning_rate": 1.5597551597004968e-06, + "loss": 0.7356, + "step": 827 + }, + { + "epoch": 2.512660028449502, + "grad_norm": 2.22249436378479, + "learning_rate": 1.5408344750555382e-06, + "loss": 0.7439, + "step": 828 + }, + { + "epoch": 2.5156946420104314, + "grad_norm": 2.2051734924316406, + "learning_rate": 1.522019667229393e-06, + "loss": 0.7271, + "step": 829 + }, + { + "epoch": 2.518729255571361, + "grad_norm": 2.0986456871032715, + "learning_rate": 1.5033109717124284e-06, + "loss": 0.7385, + "step": 830 + }, + { + "epoch": 2.52176386913229, + "grad_norm": 2.115734100341797, + "learning_rate": 1.4847086226668871e-06, + "loss": 0.7317, + "step": 831 + }, + { + "epoch": 2.5247984826932193, + "grad_norm": 2.02058482170105, + "learning_rate": 1.4662128529239572e-06, + "loss": 0.7524, + "step": 832 + }, + { + "epoch": 2.527833096254149, + "grad_norm": 2.1982414722442627, + "learning_rate": 1.4478238939808454e-06, + "loss": 0.7409, + "step": 833 + }, + { + "epoch": 2.5308677098150785, + "grad_norm": 2.1642115116119385, + "learning_rate": 1.4295419759979079e-06, + "loss": 0.7365, + "step": 834 + }, + { + "epoch": 2.5339023233760076, + "grad_norm": 2.118788003921509, + "learning_rate": 1.4113673277957395e-06, + "loss": 0.7276, + "step": 835 + }, + { + "epoch": 2.536936936936937, + "grad_norm": 2.0929243564605713, + "learning_rate": 1.393300176852327e-06, + "loss": 0.7449, + "step": 836 + }, + { + "epoch": 2.5399715504978664, + "grad_norm": 2.1339359283447266, + "learning_rate": 1.3753407493001968e-06, + "loss": 0.7241, + "step": 837 + }, + { + "epoch": 2.5430061640587955, + "grad_norm": 2.15981125831604, + "learning_rate": 1.3574892699235798e-06, + "loss": 0.7373, + "step": 838 + }, + { + "epoch": 2.546040777619725, + "grad_norm": 2.184105396270752, + "learning_rate": 1.339745962155613e-06, + "loss": 0.7283, + "step": 839 + }, + { + "epoch": 2.5490753911806543, + "grad_norm": 2.0236942768096924, + "learning_rate": 1.3221110480755306e-06, + "loss": 0.7279, + "step": 840 + }, + { + "epoch": 2.552110004741584, + "grad_norm": 2.0451905727386475, + "learning_rate": 1.3045847484058748e-06, + "loss": 0.7417, + "step": 841 + }, + { + "epoch": 2.555144618302513, + "grad_norm": 2.113330602645874, + "learning_rate": 1.287167282509767e-06, + "loss": 0.7604, + "step": 842 + }, + { + "epoch": 2.558179231863442, + "grad_norm": 2.0830495357513428, + "learning_rate": 1.2698588683881185e-06, + "loss": 0.7579, + "step": 843 + }, + { + "epoch": 2.561213845424372, + "grad_norm": 2.065068483352661, + "learning_rate": 1.252659722676941e-06, + "loss": 0.7268, + "step": 844 + }, + { + "epoch": 2.564248458985301, + "grad_norm": 2.157485008239746, + "learning_rate": 1.2355700606446119e-06, + "loss": 0.7305, + "step": 845 + }, + { + "epoch": 2.5672830725462306, + "grad_norm": 2.135056257247925, + "learning_rate": 1.2185900961891794e-06, + "loss": 0.7332, + "step": 846 + }, + { + "epoch": 2.5703176861071597, + "grad_norm": 2.1909451484680176, + "learning_rate": 1.2017200418357077e-06, + "loss": 0.7402, + "step": 847 + }, + { + "epoch": 2.5733522996680893, + "grad_norm": 2.2838058471679688, + "learning_rate": 1.184960108733586e-06, + "loss": 0.7336, + "step": 848 + }, + { + "epoch": 2.5763869132290185, + "grad_norm": 2.0348660945892334, + "learning_rate": 1.1683105066539068e-06, + "loss": 0.7367, + "step": 849 + }, + { + "epoch": 2.5794215267899476, + "grad_norm": 2.088353157043457, + "learning_rate": 1.151771443986842e-06, + "loss": 0.749, + "step": 850 + }, + { + "epoch": 2.5794215267899476, + "eval_loss": NaN, + "eval_runtime": 205.2076, + "eval_samples_per_second": 20.262, + "eval_steps_per_second": 5.068, + "step": 850 + }, + { + "epoch": 2.5824561403508772, + "grad_norm": 2.1730659008026123, + "learning_rate": 1.1353431277390125e-06, + "loss": 0.7359, + "step": 851 + }, + { + "epoch": 2.5854907539118064, + "grad_norm": 2.10697078704834, + "learning_rate": 1.1190257635309276e-06, + "loss": 0.7223, + "step": 852 + }, + { + "epoch": 2.588525367472736, + "grad_norm": 2.1502668857574463, + "learning_rate": 1.1028195555943877e-06, + "loss": 0.7265, + "step": 853 + }, + { + "epoch": 2.591559981033665, + "grad_norm": 2.0735018253326416, + "learning_rate": 1.0867247067699315e-06, + "loss": 0.7345, + "step": 854 + }, + { + "epoch": 2.5945945945945947, + "grad_norm": 2.1325292587280273, + "learning_rate": 1.0707414185043163e-06, + "loss": 0.746, + "step": 855 + }, + { + "epoch": 2.597629208155524, + "grad_norm": 2.1418328285217285, + "learning_rate": 1.0548698908479671e-06, + "loss": 0.7401, + "step": 856 + }, + { + "epoch": 2.600663821716453, + "grad_norm": 2.061847448348999, + "learning_rate": 1.0391103224524957e-06, + "loss": 0.7319, + "step": 857 + }, + { + "epoch": 2.6036984352773827, + "grad_norm": 2.1931376457214355, + "learning_rate": 1.0234629105682104e-06, + "loss": 0.7429, + "step": 858 + }, + { + "epoch": 2.6067330488383122, + "grad_norm": 2.1396734714508057, + "learning_rate": 1.0079278510416313e-06, + "loss": 0.7369, + "step": 859 + }, + { + "epoch": 2.6097676623992414, + "grad_norm": 2.2338297367095947, + "learning_rate": 9.925053383130667e-07, + "loss": 0.745, + "step": 860 + }, + { + "epoch": 2.6128022759601706, + "grad_norm": 2.120897054672241, + "learning_rate": 9.771955654141496e-07, + "loss": 0.7444, + "step": 861 + }, + { + "epoch": 2.6158368895211, + "grad_norm": 2.204875946044922, + "learning_rate": 9.619987239654405e-07, + "loss": 0.7293, + "step": 862 + }, + { + "epoch": 2.6188715030820293, + "grad_norm": 2.228181838989258, + "learning_rate": 9.469150041740338e-07, + "loss": 0.7251, + "step": 863 + }, + { + "epoch": 2.6219061166429585, + "grad_norm": 2.170053720474243, + "learning_rate": 9.319445948311534e-07, + "loss": 0.7444, + "step": 864 + }, + { + "epoch": 2.624940730203888, + "grad_norm": 2.163527011871338, + "learning_rate": 9.170876833098119e-07, + "loss": 0.7568, + "step": 865 + }, + { + "epoch": 2.6279753437648177, + "grad_norm": 2.1692657470703125, + "learning_rate": 9.023444555624572e-07, + "loss": 0.736, + "step": 866 + }, + { + "epoch": 2.631009957325747, + "grad_norm": 2.0737569332122803, + "learning_rate": 8.87715096118642e-07, + "loss": 0.7368, + "step": 867 + }, + { + "epoch": 2.634044570886676, + "grad_norm": 2.05617094039917, + "learning_rate": 8.731997880827258e-07, + "loss": 0.7334, + "step": 868 + }, + { + "epoch": 2.6370791844476056, + "grad_norm": 1.9645661115646362, + "learning_rate": 8.587987131315656e-07, + "loss": 0.7422, + "step": 869 + }, + { + "epoch": 2.6401137980085347, + "grad_norm": 2.0881567001342773, + "learning_rate": 8.445120515122552e-07, + "loss": 0.7296, + "step": 870 + }, + { + "epoch": 2.6431484115694643, + "grad_norm": 2.2319769859313965, + "learning_rate": 8.303399820398672e-07, + "loss": 0.7502, + "step": 871 + }, + { + "epoch": 2.6461830251303935, + "grad_norm": 2.1117403507232666, + "learning_rate": 8.162826820952097e-07, + "loss": 0.7319, + "step": 872 + }, + { + "epoch": 2.649217638691323, + "grad_norm": 2.056861400604248, + "learning_rate": 8.023403276226127e-07, + "loss": 0.7429, + "step": 873 + }, + { + "epoch": 2.6522522522522523, + "grad_norm": 2.140435218811035, + "learning_rate": 7.885130931277219e-07, + "loss": 0.7418, + "step": 874 + }, + { + "epoch": 2.6552868658131814, + "grad_norm": 2.11114239692688, + "learning_rate": 7.74801151675314e-07, + "loss": 0.7521, + "step": 875 + }, + { + "epoch": 2.658321479374111, + "grad_norm": 2.0694892406463623, + "learning_rate": 7.612046748871327e-07, + "loss": 0.7385, + "step": 876 + }, + { + "epoch": 2.66135609293504, + "grad_norm": 2.155579090118408, + "learning_rate": 7.477238329397419e-07, + "loss": 0.7427, + "step": 877 + }, + { + "epoch": 2.6643907064959698, + "grad_norm": 2.223646402359009, + "learning_rate": 7.343587945623908e-07, + "loss": 0.731, + "step": 878 + }, + { + "epoch": 2.667425320056899, + "grad_norm": 2.047801971435547, + "learning_rate": 7.211097270349065e-07, + "loss": 0.7426, + "step": 879 + }, + { + "epoch": 2.6704599336178285, + "grad_norm": 2.119241714477539, + "learning_rate": 7.07976796185601e-07, + "loss": 0.7407, + "step": 880 + }, + { + "epoch": 2.6734945471787577, + "grad_norm": 2.2282469272613525, + "learning_rate": 6.949601663891891e-07, + "loss": 0.7541, + "step": 881 + }, + { + "epoch": 2.676529160739687, + "grad_norm": 2.175909996032715, + "learning_rate": 6.820600005647382e-07, + "loss": 0.7461, + "step": 882 + }, + { + "epoch": 2.6795637743006164, + "grad_norm": 2.047321081161499, + "learning_rate": 6.692764601736268e-07, + "loss": 0.7197, + "step": 883 + }, + { + "epoch": 2.682598387861546, + "grad_norm": 2.0702617168426514, + "learning_rate": 6.566097052175213e-07, + "loss": 0.7656, + "step": 884 + }, + { + "epoch": 2.685633001422475, + "grad_norm": 2.1683459281921387, + "learning_rate": 6.440598942363796e-07, + "loss": 0.7521, + "step": 885 + }, + { + "epoch": 2.6886676149834043, + "grad_norm": 2.067746162414551, + "learning_rate": 6.316271843064536e-07, + "loss": 0.7531, + "step": 886 + }, + { + "epoch": 2.691702228544334, + "grad_norm": 2.1310875415802, + "learning_rate": 6.193117310383412e-07, + "loss": 0.7538, + "step": 887 + }, + { + "epoch": 2.694736842105263, + "grad_norm": 2.1403868198394775, + "learning_rate": 6.071136885750272e-07, + "loss": 0.7373, + "step": 888 + }, + { + "epoch": 2.6977714556661923, + "grad_norm": 2.049807071685791, + "learning_rate": 5.950332095899547e-07, + "loss": 0.738, + "step": 889 + }, + { + "epoch": 2.700806069227122, + "grad_norm": 2.080238103866577, + "learning_rate": 5.830704452851166e-07, + "loss": 0.7212, + "step": 890 + }, + { + "epoch": 2.7038406827880515, + "grad_norm": 2.1021158695220947, + "learning_rate": 5.71225545389158e-07, + "loss": 0.7411, + "step": 891 + }, + { + "epoch": 2.7068752963489806, + "grad_norm": 2.0948758125305176, + "learning_rate": 5.594986581555173e-07, + "loss": 0.7369, + "step": 892 + }, + { + "epoch": 2.7099099099099098, + "grad_norm": 2.1047518253326416, + "learning_rate": 5.478899303605512e-07, + "loss": 0.7289, + "step": 893 + }, + { + "epoch": 2.7129445234708394, + "grad_norm": 2.1207687854766846, + "learning_rate": 5.363995073017047e-07, + "loss": 0.7385, + "step": 894 + }, + { + "epoch": 2.7159791370317685, + "grad_norm": 2.051896572113037, + "learning_rate": 5.250275327957033e-07, + "loss": 0.7244, + "step": 895 + }, + { + "epoch": 2.719013750592698, + "grad_norm": 2.220273017883301, + "learning_rate": 5.137741491767345e-07, + "loss": 0.7522, + "step": 896 + }, + { + "epoch": 2.7220483641536273, + "grad_norm": 2.1100893020629883, + "learning_rate": 5.026394972946813e-07, + "loss": 0.7276, + "step": 897 + }, + { + "epoch": 2.725082977714557, + "grad_norm": 2.300666332244873, + "learning_rate": 4.91623716513352e-07, + "loss": 0.7404, + "step": 898 + }, + { + "epoch": 2.728117591275486, + "grad_norm": 2.2343952655792236, + "learning_rate": 4.807269447087348e-07, + "loss": 0.7474, + "step": 899 + }, + { + "epoch": 2.731152204836415, + "grad_norm": 2.1582469940185547, + "learning_rate": 4.6994931826728094e-07, + "loss": 0.7275, + "step": 900 + }, + { + "epoch": 2.731152204836415, + "eval_loss": NaN, + "eval_runtime": 204.8855, + "eval_samples_per_second": 20.294, + "eval_steps_per_second": 5.076, + "step": 900 + }, + { + "epoch": 2.734186818397345, + "grad_norm": 2.057573080062866, + "learning_rate": 4.592909720841843e-07, + "loss": 0.7351, + "step": 901 + }, + { + "epoch": 2.737221431958274, + "grad_norm": 2.320993423461914, + "learning_rate": 4.487520395617029e-07, + "loss": 0.7086, + "step": 902 + }, + { + "epoch": 2.7402560455192035, + "grad_norm": 2.2511098384857178, + "learning_rate": 4.3833265260749157e-07, + "loss": 0.723, + "step": 903 + }, + { + "epoch": 2.7432906590801327, + "grad_norm": 2.182513952255249, + "learning_rate": 4.280329416329365e-07, + "loss": 0.7196, + "step": 904 + }, + { + "epoch": 2.7463252726410623, + "grad_norm": 2.1579484939575195, + "learning_rate": 4.178530355515409e-07, + "loss": 0.7334, + "step": 905 + }, + { + "epoch": 2.7493598862019915, + "grad_norm": 2.091196060180664, + "learning_rate": 4.077930617773007e-07, + "loss": 0.7387, + "step": 906 + }, + { + "epoch": 2.7523944997629206, + "grad_norm": 2.1611766815185547, + "learning_rate": 3.97853146223105e-07, + "loss": 0.7515, + "step": 907 + }, + { + "epoch": 2.75542911332385, + "grad_norm": 2.240387201309204, + "learning_rate": 3.880334132991792e-07, + "loss": 0.7377, + "step": 908 + }, + { + "epoch": 2.7584637268847794, + "grad_norm": 2.1546630859375, + "learning_rate": 3.783339859115065e-07, + "loss": 0.7388, + "step": 909 + }, + { + "epoch": 2.761498340445709, + "grad_norm": 2.0892395973205566, + "learning_rate": 3.687549854603023e-07, + "loss": 0.7488, + "step": 910 + }, + { + "epoch": 2.764532954006638, + "grad_norm": 2.182509422302246, + "learning_rate": 3.5929653183849444e-07, + "loss": 0.736, + "step": 911 + }, + { + "epoch": 2.7675675675675677, + "grad_norm": 2.198559522628784, + "learning_rate": 3.49958743430211e-07, + "loss": 0.736, + "step": 912 + }, + { + "epoch": 2.770602181128497, + "grad_norm": 2.188133955001831, + "learning_rate": 3.4074173710931804e-07, + "loss": 0.7345, + "step": 913 + }, + { + "epoch": 2.773636794689426, + "grad_norm": 2.133892774581909, + "learning_rate": 3.3164562823793654e-07, + "loss": 0.7275, + "step": 914 + }, + { + "epoch": 2.7766714082503556, + "grad_norm": 2.159824848175049, + "learning_rate": 3.226705306650113e-07, + "loss": 0.7198, + "step": 915 + }, + { + "epoch": 2.7797060218112852, + "grad_norm": 2.1087584495544434, + "learning_rate": 3.1381655672488634e-07, + "loss": 0.7381, + "step": 916 + }, + { + "epoch": 2.7827406353722144, + "grad_norm": 2.0807528495788574, + "learning_rate": 3.050838172358883e-07, + "loss": 0.7386, + "step": 917 + }, + { + "epoch": 2.7857752489331435, + "grad_norm": 2.1473801136016846, + "learning_rate": 2.9647242149895005e-07, + "loss": 0.7336, + "step": 918 + }, + { + "epoch": 2.788809862494073, + "grad_norm": 2.219571352005005, + "learning_rate": 2.879824772962381e-07, + "loss": 0.7438, + "step": 919 + }, + { + "epoch": 2.7918444760550023, + "grad_norm": 2.144059419631958, + "learning_rate": 2.796140908898026e-07, + "loss": 0.7338, + "step": 920 + }, + { + "epoch": 2.7948790896159315, + "grad_norm": 2.1552734375, + "learning_rate": 2.7136736702025436e-07, + "loss": 0.7345, + "step": 921 + }, + { + "epoch": 2.797913703176861, + "grad_norm": 2.18730092048645, + "learning_rate": 2.632424089054419e-07, + "loss": 0.7222, + "step": 922 + }, + { + "epoch": 2.8009483167377907, + "grad_norm": 2.233304500579834, + "learning_rate": 2.552393182391677e-07, + "loss": 0.7474, + "step": 923 + }, + { + "epoch": 2.80398293029872, + "grad_norm": 2.041757583618164, + "learning_rate": 2.473581951899184e-07, + "loss": 0.7329, + "step": 924 + }, + { + "epoch": 2.807017543859649, + "grad_norm": 2.1639747619628906, + "learning_rate": 2.395991383995999e-07, + "loss": 0.748, + "step": 925 + }, + { + "epoch": 2.8100521574205786, + "grad_norm": 2.0612905025482178, + "learning_rate": 2.3196224498231447e-07, + "loss": 0.7311, + "step": 926 + }, + { + "epoch": 2.8130867709815077, + "grad_norm": 2.260965347290039, + "learning_rate": 2.2444761052313857e-07, + "loss": 0.7353, + "step": 927 + }, + { + "epoch": 2.8161213845424373, + "grad_norm": 2.1008665561676025, + "learning_rate": 2.1705532907692617e-07, + "loss": 0.7677, + "step": 928 + }, + { + "epoch": 2.8191559981033665, + "grad_norm": 2.20302414894104, + "learning_rate": 2.0978549316713615e-07, + "loss": 0.7447, + "step": 929 + }, + { + "epoch": 2.822190611664296, + "grad_norm": 2.100362539291382, + "learning_rate": 2.0263819378466888e-07, + "loss": 0.7393, + "step": 930 + }, + { + "epoch": 2.8252252252252252, + "grad_norm": 2.2228076457977295, + "learning_rate": 1.9561352038673264e-07, + "loss": 0.7501, + "step": 931 + }, + { + "epoch": 2.8282598387861544, + "grad_norm": 2.2403063774108887, + "learning_rate": 1.8871156089572018e-07, + "loss": 0.7508, + "step": 932 + }, + { + "epoch": 2.831294452347084, + "grad_norm": 2.1292290687561035, + "learning_rate": 1.8193240169810943e-07, + "loss": 0.7362, + "step": 933 + }, + { + "epoch": 2.834329065908013, + "grad_norm": 2.029907703399658, + "learning_rate": 1.752761276433801e-07, + "loss": 0.7504, + "step": 934 + }, + { + "epoch": 2.8373636794689427, + "grad_norm": 2.175046682357788, + "learning_rate": 1.6874282204295765e-07, + "loss": 0.7437, + "step": 935 + }, + { + "epoch": 2.840398293029872, + "grad_norm": 2.150054693222046, + "learning_rate": 1.623325666691644e-07, + "loss": 0.7388, + "step": 936 + }, + { + "epoch": 2.8434329065908015, + "grad_norm": 2.2034387588500977, + "learning_rate": 1.5604544175419901e-07, + "loss": 0.7451, + "step": 937 + }, + { + "epoch": 2.8464675201517307, + "grad_norm": 2.1042325496673584, + "learning_rate": 1.4988152598913063e-07, + "loss": 0.7535, + "step": 938 + }, + { + "epoch": 2.84950213371266, + "grad_norm": 2.1048696041107178, + "learning_rate": 1.4384089652291544e-07, + "loss": 0.745, + "step": 939 + }, + { + "epoch": 2.8525367472735894, + "grad_norm": 2.3105156421661377, + "learning_rate": 1.3792362896143164e-07, + "loss": 0.7233, + "step": 940 + }, + { + "epoch": 2.8555713608345186, + "grad_norm": 2.1417715549468994, + "learning_rate": 1.3212979736653142e-07, + "loss": 0.7467, + "step": 941 + }, + { + "epoch": 2.858605974395448, + "grad_norm": 2.1809537410736084, + "learning_rate": 1.2645947425511397e-07, + "loss": 0.7432, + "step": 942 + }, + { + "epoch": 2.8616405879563773, + "grad_norm": 2.2112855911254883, + "learning_rate": 1.209127305982205e-07, + "loss": 0.7487, + "step": 943 + }, + { + "epoch": 2.864675201517307, + "grad_norm": 2.059002161026001, + "learning_rate": 1.1548963582013961e-07, + "loss": 0.734, + "step": 944 + }, + { + "epoch": 2.867709815078236, + "grad_norm": 2.04377818107605, + "learning_rate": 1.1019025779754666e-07, + "loss": 0.7213, + "step": 945 + }, + { + "epoch": 2.8707444286391652, + "grad_norm": 2.0912530422210693, + "learning_rate": 1.0501466285865124e-07, + "loss": 0.7548, + "step": 946 + }, + { + "epoch": 2.873779042200095, + "grad_norm": 2.2096190452575684, + "learning_rate": 9.996291578236228e-08, + "loss": 0.7246, + "step": 947 + }, + { + "epoch": 2.8768136557610244, + "grad_norm": 2.284653663635254, + "learning_rate": 9.503507979748305e-08, + "loss": 0.7339, + "step": 948 + }, + { + "epoch": 2.8798482693219536, + "grad_norm": 2.097752332687378, + "learning_rate": 9.023121658191636e-08, + "loss": 0.7303, + "step": 949 + }, + { + "epoch": 2.8828828828828827, + "grad_norm": 2.171391010284424, + "learning_rate": 8.555138626189619e-08, + "loss": 0.7503, + "step": 950 + }, + { + "epoch": 2.8828828828828827, + "eval_loss": NaN, + "eval_runtime": 204.3654, + "eval_samples_per_second": 20.346, + "eval_steps_per_second": 5.089, + "step": 950 + }, + { + "epoch": 2.8859174964438123, + "grad_norm": 2.1366286277770996, + "learning_rate": 8.099564741123167e-08, + "loss": 0.7241, + "step": 951 + }, + { + "epoch": 2.8889521100047415, + "grad_norm": 2.0811874866485596, + "learning_rate": 7.656405705057434e-08, + "loss": 0.7504, + "step": 952 + }, + { + "epoch": 2.8919867235656707, + "grad_norm": 2.080226421356201, + "learning_rate": 7.225667064670761e-08, + "loss": 0.7355, + "step": 953 + }, + { + "epoch": 2.8950213371266003, + "grad_norm": 2.200331211090088, + "learning_rate": 6.807354211184613e-08, + "loss": 0.7427, + "step": 954 + }, + { + "epoch": 2.89805595068753, + "grad_norm": 2.038591146469116, + "learning_rate": 6.401472380297091e-08, + "loss": 0.7379, + "step": 955 + }, + { + "epoch": 2.901090564248459, + "grad_norm": 2.1141886711120605, + "learning_rate": 6.008026652116306e-08, + "loss": 0.7376, + "step": 956 + }, + { + "epoch": 2.904125177809388, + "grad_norm": 2.0803427696228027, + "learning_rate": 5.6270219510975445e-08, + "loss": 0.7424, + "step": 957 + }, + { + "epoch": 2.9071597913703178, + "grad_norm": 2.0900638103485107, + "learning_rate": 5.258463045981432e-08, + "loss": 0.7243, + "step": 958 + }, + { + "epoch": 2.910194404931247, + "grad_norm": 2.133312463760376, + "learning_rate": 4.902354549733979e-08, + "loss": 0.7257, + "step": 959 + }, + { + "epoch": 2.9132290184921765, + "grad_norm": 2.2082343101501465, + "learning_rate": 4.5587009194894005e-08, + "loss": 0.7346, + "step": 960 + }, + { + "epoch": 2.9162636320531057, + "grad_norm": 2.0872416496276855, + "learning_rate": 4.227506456493835e-08, + "loss": 0.7437, + "step": 961 + }, + { + "epoch": 2.9192982456140353, + "grad_norm": 2.136866807937622, + "learning_rate": 3.908775306051604e-08, + "loss": 0.7362, + "step": 962 + }, + { + "epoch": 2.9223328591749644, + "grad_norm": 2.125523805618286, + "learning_rate": 3.602511457473479e-08, + "loss": 0.7298, + "step": 963 + }, + { + "epoch": 2.9253674727358936, + "grad_norm": 2.055690050125122, + "learning_rate": 3.3087187440268284e-08, + "loss": 0.7285, + "step": 964 + }, + { + "epoch": 2.928402086296823, + "grad_norm": 2.26811146736145, + "learning_rate": 3.027400842887218e-08, + "loss": 0.732, + "step": 965 + }, + { + "epoch": 2.9314366998577523, + "grad_norm": 2.0984137058258057, + "learning_rate": 2.758561275092886e-08, + "loss": 0.7238, + "step": 966 + }, + { + "epoch": 2.934471313418682, + "grad_norm": 2.117258310317993, + "learning_rate": 2.5022034055003363e-08, + "loss": 0.7604, + "step": 967 + }, + { + "epoch": 2.937505926979611, + "grad_norm": 2.135441541671753, + "learning_rate": 2.2583304427421516e-08, + "loss": 0.7358, + "step": 968 + }, + { + "epoch": 2.9405405405405407, + "grad_norm": 2.1416280269622803, + "learning_rate": 2.0269454391874665e-08, + "loss": 0.7436, + "step": 969 + }, + { + "epoch": 2.94357515410147, + "grad_norm": 2.1733927726745605, + "learning_rate": 1.80805129090289e-08, + "loss": 0.7308, + "step": 970 + }, + { + "epoch": 2.946609767662399, + "grad_norm": 2.1229727268218994, + "learning_rate": 1.6016507376169776e-08, + "loss": 0.7577, + "step": 971 + }, + { + "epoch": 2.9496443812233286, + "grad_norm": 2.283834218978882, + "learning_rate": 1.4077463626852584e-08, + "loss": 0.748, + "step": 972 + }, + { + "epoch": 2.952678994784258, + "grad_norm": 2.205937623977661, + "learning_rate": 1.2263405930585947e-08, + "loss": 0.7335, + "step": 973 + }, + { + "epoch": 2.9557136083451874, + "grad_norm": 2.111551284790039, + "learning_rate": 1.0574356992525403e-08, + "loss": 0.7295, + "step": 974 + }, + { + "epoch": 2.9587482219061165, + "grad_norm": 2.058469295501709, + "learning_rate": 9.010337953185843e-09, + "loss": 0.7547, + "step": 975 + }, + { + "epoch": 2.961782835467046, + "grad_norm": 2.1459245681762695, + "learning_rate": 7.571368388181732e-09, + "loss": 0.7368, + "step": 976 + }, + { + "epoch": 2.9648174490279753, + "grad_norm": 2.1419780254364014, + "learning_rate": 6.257466307980631e-09, + "loss": 0.7364, + "step": 977 + }, + { + "epoch": 2.9678520625889044, + "grad_norm": 2.1368930339813232, + "learning_rate": 5.068648157675604e-09, + "loss": 0.7243, + "step": 978 + }, + { + "epoch": 2.970886676149834, + "grad_norm": 2.072345733642578, + "learning_rate": 4.00492881678427e-09, + "loss": 0.7229, + "step": 979 + }, + { + "epoch": 2.9739212897107636, + "grad_norm": 2.173828363418579, + "learning_rate": 3.0663215990534013e-09, + "loss": 0.711, + "step": 980 + }, + { + "epoch": 2.976955903271693, + "grad_norm": 2.117293357849121, + "learning_rate": 2.2528382523057115e-09, + "loss": 0.7648, + "step": 981 + }, + { + "epoch": 2.979990516832622, + "grad_norm": 2.126967191696167, + "learning_rate": 1.564488958279986e-09, + "loss": 0.7359, + "step": 982 + }, + { + "epoch": 2.9830251303935515, + "grad_norm": 2.086338520050049, + "learning_rate": 1.0012823325111776e-09, + "loss": 0.7575, + "step": 983 + }, + { + "epoch": 2.9860597439544807, + "grad_norm": 2.0174200534820557, + "learning_rate": 5.632254242204926e-10, + "loss": 0.7324, + "step": 984 + }, + { + "epoch": 2.9890943575154103, + "grad_norm": 2.0716025829315186, + "learning_rate": 2.503237162254646e-10, + "loss": 0.7482, + "step": 985 + }, + { + "epoch": 2.9921289710763395, + "grad_norm": 2.0842347145080566, + "learning_rate": 6.258112487667056e-11, + "loss": 0.7334, + "step": 986 + }, + { + "epoch": 2.995163584637269, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.0, + "loss": 0.7299, + "step": 987 + }, + { + "epoch": 2.995163584637269, + "step": 987, + "total_flos": 4.415483185831556e+19, + "train_loss": 0.7698779804849093, + "train_runtime": 131039.1076, + "train_samples_per_second": 3.863, + "train_steps_per_second": 0.008 + } + ], + "logging_steps": 1.0, + "max_steps": 987, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 4.415483185831556e+19, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}