chat_3 / trainer_state.json
ciyazzzk
commit from ciyazhou
c330bda
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.995163584637269,
"eval_steps": 50,
"global_step": 987,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0030346135609293505,
"grad_norm": 5.274968147277832,
"learning_rate": 2.0202020202020205e-07,
"loss": 0.8522,
"step": 1
},
{
"epoch": 0.006069227121858701,
"grad_norm": 5.754838466644287,
"learning_rate": 4.040404040404041e-07,
"loss": 0.854,
"step": 2
},
{
"epoch": 0.009103840682788052,
"grad_norm": 5.632802963256836,
"learning_rate": 6.060606060606061e-07,
"loss": 0.873,
"step": 3
},
{
"epoch": 0.012138454243717402,
"grad_norm": 5.394160747528076,
"learning_rate": 8.080808080808082e-07,
"loss": 0.8563,
"step": 4
},
{
"epoch": 0.015173067804646752,
"grad_norm": 4.5878424644470215,
"learning_rate": 1.01010101010101e-06,
"loss": 0.8425,
"step": 5
},
{
"epoch": 0.018207681365576104,
"grad_norm": 4.280163764953613,
"learning_rate": 1.2121212121212122e-06,
"loss": 0.8569,
"step": 6
},
{
"epoch": 0.021242294926505454,
"grad_norm": 3.5766515731811523,
"learning_rate": 1.4141414141414143e-06,
"loss": 0.8309,
"step": 7
},
{
"epoch": 0.024276908487434804,
"grad_norm": 4.070709228515625,
"learning_rate": 1.6161616161616164e-06,
"loss": 0.825,
"step": 8
},
{
"epoch": 0.027311522048364154,
"grad_norm": 4.395944595336914,
"learning_rate": 1.8181818181818183e-06,
"loss": 0.8419,
"step": 9
},
{
"epoch": 0.030346135609293504,
"grad_norm": 3.765310764312744,
"learning_rate": 2.02020202020202e-06,
"loss": 0.8426,
"step": 10
},
{
"epoch": 0.03338074917022286,
"grad_norm": 2.906996488571167,
"learning_rate": 2.222222222222222e-06,
"loss": 0.8263,
"step": 11
},
{
"epoch": 0.03641536273115221,
"grad_norm": 2.8421216011047363,
"learning_rate": 2.4242424242424244e-06,
"loss": 0.8242,
"step": 12
},
{
"epoch": 0.03944997629208156,
"grad_norm": 2.6818017959594727,
"learning_rate": 2.6262626262626267e-06,
"loss": 0.8206,
"step": 13
},
{
"epoch": 0.04248458985301091,
"grad_norm": 2.8597841262817383,
"learning_rate": 2.8282828282828286e-06,
"loss": 0.8201,
"step": 14
},
{
"epoch": 0.04551920341394026,
"grad_norm": 2.673666000366211,
"learning_rate": 3.0303030303030305e-06,
"loss": 0.8215,
"step": 15
},
{
"epoch": 0.04855381697486961,
"grad_norm": 2.636281728744507,
"learning_rate": 3.232323232323233e-06,
"loss": 0.8037,
"step": 16
},
{
"epoch": 0.05158843053579896,
"grad_norm": 2.502042770385742,
"learning_rate": 3.4343434343434347e-06,
"loss": 0.8162,
"step": 17
},
{
"epoch": 0.05462304409672831,
"grad_norm": 2.562307834625244,
"learning_rate": 3.6363636363636366e-06,
"loss": 0.8066,
"step": 18
},
{
"epoch": 0.05765765765765766,
"grad_norm": 2.6667816638946533,
"learning_rate": 3.8383838383838385e-06,
"loss": 0.8197,
"step": 19
},
{
"epoch": 0.06069227121858701,
"grad_norm": 2.293628215789795,
"learning_rate": 4.04040404040404e-06,
"loss": 0.8092,
"step": 20
},
{
"epoch": 0.06372688477951635,
"grad_norm": 2.4843485355377197,
"learning_rate": 4.242424242424243e-06,
"loss": 0.8082,
"step": 21
},
{
"epoch": 0.06676149834044572,
"grad_norm": 2.5178024768829346,
"learning_rate": 4.444444444444444e-06,
"loss": 0.8165,
"step": 22
},
{
"epoch": 0.06979611190137507,
"grad_norm": 2.6686830520629883,
"learning_rate": 4.646464646464647e-06,
"loss": 0.773,
"step": 23
},
{
"epoch": 0.07283072546230442,
"grad_norm": 2.2187390327453613,
"learning_rate": 4.848484848484849e-06,
"loss": 0.8026,
"step": 24
},
{
"epoch": 0.07586533902323377,
"grad_norm": 2.4837453365325928,
"learning_rate": 5.0505050505050515e-06,
"loss": 0.7962,
"step": 25
},
{
"epoch": 0.07889995258416312,
"grad_norm": 2.3953895568847656,
"learning_rate": 5.252525252525253e-06,
"loss": 0.7902,
"step": 26
},
{
"epoch": 0.08193456614509247,
"grad_norm": 2.384993553161621,
"learning_rate": 5.4545454545454545e-06,
"loss": 0.8133,
"step": 27
},
{
"epoch": 0.08496917970602182,
"grad_norm": 2.5588831901550293,
"learning_rate": 5.656565656565657e-06,
"loss": 0.8043,
"step": 28
},
{
"epoch": 0.08800379326695117,
"grad_norm": 2.254196882247925,
"learning_rate": 5.858585858585859e-06,
"loss": 0.7967,
"step": 29
},
{
"epoch": 0.09103840682788052,
"grad_norm": 2.2686617374420166,
"learning_rate": 6.060606060606061e-06,
"loss": 0.8036,
"step": 30
},
{
"epoch": 0.09407302038880987,
"grad_norm": 2.3785650730133057,
"learning_rate": 6.262626262626264e-06,
"loss": 0.799,
"step": 31
},
{
"epoch": 0.09710763394973922,
"grad_norm": 2.4035139083862305,
"learning_rate": 6.464646464646466e-06,
"loss": 0.8086,
"step": 32
},
{
"epoch": 0.10014224751066857,
"grad_norm": 2.2935521602630615,
"learning_rate": 6.666666666666667e-06,
"loss": 0.7842,
"step": 33
},
{
"epoch": 0.10317686107159792,
"grad_norm": 2.3163795471191406,
"learning_rate": 6.868686868686869e-06,
"loss": 0.8041,
"step": 34
},
{
"epoch": 0.10621147463252727,
"grad_norm": 2.3953793048858643,
"learning_rate": 7.070707070707071e-06,
"loss": 0.7929,
"step": 35
},
{
"epoch": 0.10924608819345662,
"grad_norm": 2.4116768836975098,
"learning_rate": 7.272727272727273e-06,
"loss": 0.7943,
"step": 36
},
{
"epoch": 0.11228070175438597,
"grad_norm": 2.5221264362335205,
"learning_rate": 7.474747474747476e-06,
"loss": 0.8036,
"step": 37
},
{
"epoch": 0.11531531531531532,
"grad_norm": 2.4730491638183594,
"learning_rate": 7.676767676767677e-06,
"loss": 0.8063,
"step": 38
},
{
"epoch": 0.11834992887624467,
"grad_norm": 2.1653268337249756,
"learning_rate": 7.87878787878788e-06,
"loss": 0.8009,
"step": 39
},
{
"epoch": 0.12138454243717402,
"grad_norm": 2.266632080078125,
"learning_rate": 8.08080808080808e-06,
"loss": 0.7825,
"step": 40
},
{
"epoch": 0.12441915599810337,
"grad_norm": 2.2630982398986816,
"learning_rate": 8.282828282828283e-06,
"loss": 0.8004,
"step": 41
},
{
"epoch": 0.1274537695590327,
"grad_norm": 2.403125762939453,
"learning_rate": 8.484848484848486e-06,
"loss": 0.7911,
"step": 42
},
{
"epoch": 0.13048838311996205,
"grad_norm": 2.4483842849731445,
"learning_rate": 8.686868686868687e-06,
"loss": 0.7861,
"step": 43
},
{
"epoch": 0.13352299668089143,
"grad_norm": 2.346780776977539,
"learning_rate": 8.888888888888888e-06,
"loss": 0.7969,
"step": 44
},
{
"epoch": 0.13655761024182078,
"grad_norm": 2.3152859210968018,
"learning_rate": 9.090909090909091e-06,
"loss": 0.7842,
"step": 45
},
{
"epoch": 0.13959222380275013,
"grad_norm": 2.2333388328552246,
"learning_rate": 9.292929292929294e-06,
"loss": 0.7886,
"step": 46
},
{
"epoch": 0.14262683736367948,
"grad_norm": 2.360898017883301,
"learning_rate": 9.494949494949497e-06,
"loss": 0.8146,
"step": 47
},
{
"epoch": 0.14566145092460883,
"grad_norm": 2.5461983680725098,
"learning_rate": 9.696969696969698e-06,
"loss": 0.8005,
"step": 48
},
{
"epoch": 0.14869606448553818,
"grad_norm": 2.4684877395629883,
"learning_rate": 9.8989898989899e-06,
"loss": 0.8031,
"step": 49
},
{
"epoch": 0.15173067804646753,
"grad_norm": 2.3025074005126953,
"learning_rate": 1.0101010101010103e-05,
"loss": 0.7828,
"step": 50
},
{
"epoch": 0.15173067804646753,
"eval_loss": NaN,
"eval_runtime": 233.599,
"eval_samples_per_second": 17.8,
"eval_steps_per_second": 4.452,
"step": 50
},
{
"epoch": 0.15476529160739688,
"grad_norm": 2.4276065826416016,
"learning_rate": 1.0303030303030304e-05,
"loss": 0.7801,
"step": 51
},
{
"epoch": 0.15779990516832623,
"grad_norm": 2.279287815093994,
"learning_rate": 1.0505050505050507e-05,
"loss": 0.8055,
"step": 52
},
{
"epoch": 0.16083451872925558,
"grad_norm": 2.7459404468536377,
"learning_rate": 1.0707070707070708e-05,
"loss": 0.7864,
"step": 53
},
{
"epoch": 0.16386913229018493,
"grad_norm": 2.3488147258758545,
"learning_rate": 1.0909090909090909e-05,
"loss": 0.7869,
"step": 54
},
{
"epoch": 0.16690374585111428,
"grad_norm": 2.547166109085083,
"learning_rate": 1.1111111111111113e-05,
"loss": 0.7869,
"step": 55
},
{
"epoch": 0.16993835941204363,
"grad_norm": 2.4360105991363525,
"learning_rate": 1.1313131313131314e-05,
"loss": 0.7883,
"step": 56
},
{
"epoch": 0.17297297297297298,
"grad_norm": 2.8864669799804688,
"learning_rate": 1.1515151515151517e-05,
"loss": 0.7927,
"step": 57
},
{
"epoch": 0.17600758653390233,
"grad_norm": 2.3167998790740967,
"learning_rate": 1.1717171717171718e-05,
"loss": 0.7999,
"step": 58
},
{
"epoch": 0.17904220009483168,
"grad_norm": 3.075058698654175,
"learning_rate": 1.191919191919192e-05,
"loss": 0.7888,
"step": 59
},
{
"epoch": 0.18207681365576103,
"grad_norm": 2.38899827003479,
"learning_rate": 1.2121212121212122e-05,
"loss": 0.787,
"step": 60
},
{
"epoch": 0.18511142721669038,
"grad_norm": 3.121044874191284,
"learning_rate": 1.2323232323232323e-05,
"loss": 0.8053,
"step": 61
},
{
"epoch": 0.18814604077761973,
"grad_norm": 2.580725908279419,
"learning_rate": 1.2525252525252527e-05,
"loss": 0.7819,
"step": 62
},
{
"epoch": 0.19118065433854908,
"grad_norm": 3.1028575897216797,
"learning_rate": 1.2727272727272728e-05,
"loss": 0.7886,
"step": 63
},
{
"epoch": 0.19421526789947843,
"grad_norm": 2.6575424671173096,
"learning_rate": 1.2929292929292931e-05,
"loss": 0.7904,
"step": 64
},
{
"epoch": 0.19724988146040778,
"grad_norm": 2.8595755100250244,
"learning_rate": 1.3131313131313132e-05,
"loss": 0.7955,
"step": 65
},
{
"epoch": 0.20028449502133713,
"grad_norm": 2.235410451889038,
"learning_rate": 1.3333333333333333e-05,
"loss": 0.796,
"step": 66
},
{
"epoch": 0.20331910858226648,
"grad_norm": 2.7857375144958496,
"learning_rate": 1.3535353535353538e-05,
"loss": 0.8008,
"step": 67
},
{
"epoch": 0.20635372214319583,
"grad_norm": 2.310331106185913,
"learning_rate": 1.3737373737373739e-05,
"loss": 0.7802,
"step": 68
},
{
"epoch": 0.20938833570412518,
"grad_norm": 2.8039300441741943,
"learning_rate": 1.3939393939393942e-05,
"loss": 0.7946,
"step": 69
},
{
"epoch": 0.21242294926505453,
"grad_norm": 2.345369815826416,
"learning_rate": 1.4141414141414143e-05,
"loss": 0.7835,
"step": 70
},
{
"epoch": 0.21545756282598388,
"grad_norm": 2.8384764194488525,
"learning_rate": 1.4343434343434344e-05,
"loss": 0.7855,
"step": 71
},
{
"epoch": 0.21849217638691323,
"grad_norm": 2.371159076690674,
"learning_rate": 1.4545454545454546e-05,
"loss": 0.7689,
"step": 72
},
{
"epoch": 0.22152678994784258,
"grad_norm": 2.9578475952148438,
"learning_rate": 1.4747474747474747e-05,
"loss": 0.7918,
"step": 73
},
{
"epoch": 0.22456140350877193,
"grad_norm": 2.8900325298309326,
"learning_rate": 1.4949494949494952e-05,
"loss": 0.7746,
"step": 74
},
{
"epoch": 0.22759601706970128,
"grad_norm": 2.4469008445739746,
"learning_rate": 1.5151515151515153e-05,
"loss": 0.7919,
"step": 75
},
{
"epoch": 0.23063063063063063,
"grad_norm": 2.410154104232788,
"learning_rate": 1.5353535353535354e-05,
"loss": 0.7987,
"step": 76
},
{
"epoch": 0.23366524419155998,
"grad_norm": 2.599086284637451,
"learning_rate": 1.555555555555556e-05,
"loss": 0.7714,
"step": 77
},
{
"epoch": 0.23669985775248933,
"grad_norm": 2.4329092502593994,
"learning_rate": 1.575757575757576e-05,
"loss": 0.8041,
"step": 78
},
{
"epoch": 0.23973447131341868,
"grad_norm": 2.7239413261413574,
"learning_rate": 1.595959595959596e-05,
"loss": 0.7941,
"step": 79
},
{
"epoch": 0.24276908487434803,
"grad_norm": 2.8684301376342773,
"learning_rate": 1.616161616161616e-05,
"loss": 0.8076,
"step": 80
},
{
"epoch": 0.24580369843527738,
"grad_norm": 3.004608631134033,
"learning_rate": 1.6363636363636366e-05,
"loss": 0.7893,
"step": 81
},
{
"epoch": 0.24883831199620673,
"grad_norm": 2.6391353607177734,
"learning_rate": 1.6565656565656567e-05,
"loss": 0.8116,
"step": 82
},
{
"epoch": 0.2518729255571361,
"grad_norm": 3.3293023109436035,
"learning_rate": 1.6767676767676768e-05,
"loss": 0.7952,
"step": 83
},
{
"epoch": 0.2549075391180654,
"grad_norm": 2.5788087844848633,
"learning_rate": 1.6969696969696972e-05,
"loss": 0.7905,
"step": 84
},
{
"epoch": 0.2579421526789948,
"grad_norm": 3.4907588958740234,
"learning_rate": 1.7171717171717173e-05,
"loss": 0.7903,
"step": 85
},
{
"epoch": 0.2609767662399241,
"grad_norm": 2.8050403594970703,
"learning_rate": 1.7373737373737375e-05,
"loss": 0.7813,
"step": 86
},
{
"epoch": 0.2640113798008535,
"grad_norm": 3.180819034576416,
"learning_rate": 1.7575757575757576e-05,
"loss": 0.7797,
"step": 87
},
{
"epoch": 0.26704599336178286,
"grad_norm": 2.614614486694336,
"learning_rate": 1.7777777777777777e-05,
"loss": 0.7903,
"step": 88
},
{
"epoch": 0.2700806069227122,
"grad_norm": 3.212218761444092,
"learning_rate": 1.797979797979798e-05,
"loss": 0.7927,
"step": 89
},
{
"epoch": 0.27311522048364156,
"grad_norm": 2.6879336833953857,
"learning_rate": 1.8181818181818182e-05,
"loss": 0.8099,
"step": 90
},
{
"epoch": 0.2761498340445709,
"grad_norm": 3.2700448036193848,
"learning_rate": 1.8383838383838387e-05,
"loss": 0.794,
"step": 91
},
{
"epoch": 0.27918444760550026,
"grad_norm": 3.1249783039093018,
"learning_rate": 1.8585858585858588e-05,
"loss": 0.7807,
"step": 92
},
{
"epoch": 0.2822190611664296,
"grad_norm": 2.4789459705352783,
"learning_rate": 1.8787878787878792e-05,
"loss": 0.7829,
"step": 93
},
{
"epoch": 0.28525367472735896,
"grad_norm": 2.624887466430664,
"learning_rate": 1.8989898989898993e-05,
"loss": 0.8043,
"step": 94
},
{
"epoch": 0.2882882882882883,
"grad_norm": 2.3336539268493652,
"learning_rate": 1.9191919191919194e-05,
"loss": 0.7827,
"step": 95
},
{
"epoch": 0.29132290184921766,
"grad_norm": 2.5490119457244873,
"learning_rate": 1.9393939393939395e-05,
"loss": 0.8072,
"step": 96
},
{
"epoch": 0.294357515410147,
"grad_norm": 2.5659990310668945,
"learning_rate": 1.9595959595959596e-05,
"loss": 0.7986,
"step": 97
},
{
"epoch": 0.29739212897107636,
"grad_norm": 2.3554656505584717,
"learning_rate": 1.97979797979798e-05,
"loss": 0.8059,
"step": 98
},
{
"epoch": 0.3004267425320057,
"grad_norm": 2.4364328384399414,
"learning_rate": 2e-05,
"loss": 0.7874,
"step": 99
},
{
"epoch": 0.30346135609293506,
"grad_norm": 2.3030965328216553,
"learning_rate": 1.9999937418875125e-05,
"loss": 0.7854,
"step": 100
},
{
"epoch": 0.30346135609293506,
"eval_loss": NaN,
"eval_runtime": 204.7053,
"eval_samples_per_second": 20.312,
"eval_steps_per_second": 5.08,
"step": 100
},
{
"epoch": 0.3064959696538644,
"grad_norm": 2.318124294281006,
"learning_rate": 1.9999749676283775e-05,
"loss": 0.7964,
"step": 101
},
{
"epoch": 0.30953058321479376,
"grad_norm": 2.2906177043914795,
"learning_rate": 1.9999436774575783e-05,
"loss": 0.8049,
"step": 102
},
{
"epoch": 0.3125651967757231,
"grad_norm": 2.2947778701782227,
"learning_rate": 1.999899871766749e-05,
"loss": 0.7881,
"step": 103
},
{
"epoch": 0.31559981033665246,
"grad_norm": 2.548788547515869,
"learning_rate": 1.999843551104172e-05,
"loss": 0.803,
"step": 104
},
{
"epoch": 0.3186344238975818,
"grad_norm": 2.3735954761505127,
"learning_rate": 1.9997747161747696e-05,
"loss": 0.8044,
"step": 105
},
{
"epoch": 0.32166903745851116,
"grad_norm": 2.620364189147949,
"learning_rate": 1.9996933678400948e-05,
"loss": 0.7806,
"step": 106
},
{
"epoch": 0.3247036510194405,
"grad_norm": 2.3897509574890137,
"learning_rate": 1.999599507118322e-05,
"loss": 0.7862,
"step": 107
},
{
"epoch": 0.32773826458036986,
"grad_norm": 2.6068966388702393,
"learning_rate": 1.9994931351842327e-05,
"loss": 0.8051,
"step": 108
},
{
"epoch": 0.3307728781412992,
"grad_norm": 2.560683488845825,
"learning_rate": 1.999374253369202e-05,
"loss": 0.7956,
"step": 109
},
{
"epoch": 0.33380749170222856,
"grad_norm": 2.1629045009613037,
"learning_rate": 1.999242863161182e-05,
"loss": 0.8074,
"step": 110
},
{
"epoch": 0.3368421052631579,
"grad_norm": 2.653731346130371,
"learning_rate": 1.999098966204682e-05,
"loss": 0.8071,
"step": 111
},
{
"epoch": 0.33987671882408726,
"grad_norm": 2.484330177307129,
"learning_rate": 1.9989425643007475e-05,
"loss": 0.8135,
"step": 112
},
{
"epoch": 0.3429113323850166,
"grad_norm": 2.3897571563720703,
"learning_rate": 1.9987736594069417e-05,
"loss": 0.8066,
"step": 113
},
{
"epoch": 0.34594594594594597,
"grad_norm": 2.653904676437378,
"learning_rate": 1.998592253637315e-05,
"loss": 0.7908,
"step": 114
},
{
"epoch": 0.3489805595068753,
"grad_norm": 2.2755186557769775,
"learning_rate": 1.9983983492623832e-05,
"loss": 0.7824,
"step": 115
},
{
"epoch": 0.35201517306780467,
"grad_norm": 2.260126829147339,
"learning_rate": 1.9981919487090973e-05,
"loss": 0.7916,
"step": 116
},
{
"epoch": 0.355049786628734,
"grad_norm": 2.6538352966308594,
"learning_rate": 1.9979730545608128e-05,
"loss": 0.7927,
"step": 117
},
{
"epoch": 0.35808440018966337,
"grad_norm": 2.2571558952331543,
"learning_rate": 1.9977416695572577e-05,
"loss": 0.7826,
"step": 118
},
{
"epoch": 0.3611190137505927,
"grad_norm": 2.5154271125793457,
"learning_rate": 1.9974977965945e-05,
"loss": 0.807,
"step": 119
},
{
"epoch": 0.36415362731152207,
"grad_norm": 2.4908971786499023,
"learning_rate": 1.9972414387249074e-05,
"loss": 0.7831,
"step": 120
},
{
"epoch": 0.3671882408724514,
"grad_norm": 2.3925859928131104,
"learning_rate": 1.996972599157113e-05,
"loss": 0.7844,
"step": 121
},
{
"epoch": 0.37022285443338077,
"grad_norm": 2.3527793884277344,
"learning_rate": 1.9966912812559733e-05,
"loss": 0.7921,
"step": 122
},
{
"epoch": 0.3732574679943101,
"grad_norm": 2.2694365978240967,
"learning_rate": 1.9963974885425267e-05,
"loss": 0.7816,
"step": 123
},
{
"epoch": 0.37629208155523947,
"grad_norm": 2.436676025390625,
"learning_rate": 1.9960912246939486e-05,
"loss": 0.7782,
"step": 124
},
{
"epoch": 0.3793266951161688,
"grad_norm": 2.3587653636932373,
"learning_rate": 1.9957724935435065e-05,
"loss": 0.8024,
"step": 125
},
{
"epoch": 0.38236130867709817,
"grad_norm": 2.3145172595977783,
"learning_rate": 1.9954412990805107e-05,
"loss": 0.8115,
"step": 126
},
{
"epoch": 0.3853959222380275,
"grad_norm": 2.271946430206299,
"learning_rate": 1.995097645450266e-05,
"loss": 0.7975,
"step": 127
},
{
"epoch": 0.38843053579895687,
"grad_norm": 2.2533860206604004,
"learning_rate": 1.994741536954019e-05,
"loss": 0.8187,
"step": 128
},
{
"epoch": 0.3914651493598862,
"grad_norm": 2.5504581928253174,
"learning_rate": 1.994372978048903e-05,
"loss": 0.7913,
"step": 129
},
{
"epoch": 0.39449976292081557,
"grad_norm": 2.3467888832092285,
"learning_rate": 1.993991973347884e-05,
"loss": 0.7955,
"step": 130
},
{
"epoch": 0.3975343764817449,
"grad_norm": 2.356933832168579,
"learning_rate": 1.9935985276197033e-05,
"loss": 0.808,
"step": 131
},
{
"epoch": 0.40056899004267427,
"grad_norm": 2.608546257019043,
"learning_rate": 1.9931926457888155e-05,
"loss": 0.785,
"step": 132
},
{
"epoch": 0.4036036036036036,
"grad_norm": 2.2730495929718018,
"learning_rate": 1.9927743329353295e-05,
"loss": 0.79,
"step": 133
},
{
"epoch": 0.40663821716453297,
"grad_norm": 2.2720224857330322,
"learning_rate": 1.992343594294943e-05,
"loss": 0.8084,
"step": 134
},
{
"epoch": 0.4096728307254623,
"grad_norm": 2.2325122356414795,
"learning_rate": 1.9919004352588768e-05,
"loss": 0.8008,
"step": 135
},
{
"epoch": 0.41270744428639167,
"grad_norm": 2.4259414672851562,
"learning_rate": 1.9914448613738107e-05,
"loss": 0.7827,
"step": 136
},
{
"epoch": 0.415742057847321,
"grad_norm": 2.2663819789886475,
"learning_rate": 1.9909768783418086e-05,
"loss": 0.8059,
"step": 137
},
{
"epoch": 0.41877667140825037,
"grad_norm": 2.910830020904541,
"learning_rate": 1.990496492020252e-05,
"loss": 0.8159,
"step": 138
},
{
"epoch": 0.4218112849691797,
"grad_norm": 2.4485902786254883,
"learning_rate": 1.9900037084217637e-05,
"loss": 0.7921,
"step": 139
},
{
"epoch": 0.42484589853010907,
"grad_norm": 2.3199424743652344,
"learning_rate": 1.989498533714135e-05,
"loss": 0.8006,
"step": 140
},
{
"epoch": 0.4278805120910384,
"grad_norm": 2.529900550842285,
"learning_rate": 1.9889809742202454e-05,
"loss": 0.7805,
"step": 141
},
{
"epoch": 0.43091512565196777,
"grad_norm": 2.5135438442230225,
"learning_rate": 1.988451036417986e-05,
"loss": 0.8088,
"step": 142
},
{
"epoch": 0.4339497392128971,
"grad_norm": 2.359200954437256,
"learning_rate": 1.9879087269401782e-05,
"loss": 0.7963,
"step": 143
},
{
"epoch": 0.43698435277382647,
"grad_norm": 2.307164192199707,
"learning_rate": 1.9873540525744888e-05,
"loss": 0.79,
"step": 144
},
{
"epoch": 0.4400189663347558,
"grad_norm": 2.341834545135498,
"learning_rate": 1.986787020263347e-05,
"loss": 0.7955,
"step": 145
},
{
"epoch": 0.44305357989568517,
"grad_norm": 2.409449338912964,
"learning_rate": 1.986207637103857e-05,
"loss": 0.7761,
"step": 146
},
{
"epoch": 0.4460881934566145,
"grad_norm": 2.3119232654571533,
"learning_rate": 1.9856159103477085e-05,
"loss": 0.7992,
"step": 147
},
{
"epoch": 0.44912280701754387,
"grad_norm": 2.3376715183258057,
"learning_rate": 1.9850118474010873e-05,
"loss": 0.784,
"step": 148
},
{
"epoch": 0.4521574205784732,
"grad_norm": 2.387392997741699,
"learning_rate": 1.98439545582458e-05,
"loss": 0.7806,
"step": 149
},
{
"epoch": 0.45519203413940257,
"grad_norm": 2.2223150730133057,
"learning_rate": 1.983766743333084e-05,
"loss": 0.7914,
"step": 150
},
{
"epoch": 0.45519203413940257,
"eval_loss": NaN,
"eval_runtime": 205.0592,
"eval_samples_per_second": 20.277,
"eval_steps_per_second": 5.072,
"step": 150
},
{
"epoch": 0.4582266477003319,
"grad_norm": 2.727046012878418,
"learning_rate": 1.9831257177957045e-05,
"loss": 0.793,
"step": 151
},
{
"epoch": 0.46126126126126127,
"grad_norm": 2.266788959503174,
"learning_rate": 1.9824723872356623e-05,
"loss": 0.8127,
"step": 152
},
{
"epoch": 0.4642958748221906,
"grad_norm": 2.530904531478882,
"learning_rate": 1.9818067598301894e-05,
"loss": 0.7973,
"step": 153
},
{
"epoch": 0.46733048838311997,
"grad_norm": 2.3305609226226807,
"learning_rate": 1.981128843910428e-05,
"loss": 0.7961,
"step": 154
},
{
"epoch": 0.4703651019440493,
"grad_norm": 2.1240079402923584,
"learning_rate": 1.9804386479613268e-05,
"loss": 0.8093,
"step": 155
},
{
"epoch": 0.47339971550497867,
"grad_norm": 2.625185489654541,
"learning_rate": 1.9797361806215335e-05,
"loss": 0.7988,
"step": 156
},
{
"epoch": 0.476434329065908,
"grad_norm": 2.4407734870910645,
"learning_rate": 1.9790214506832868e-05,
"loss": 0.8166,
"step": 157
},
{
"epoch": 0.47946894262683737,
"grad_norm": 2.4013476371765137,
"learning_rate": 1.9782944670923075e-05,
"loss": 0.7935,
"step": 158
},
{
"epoch": 0.4825035561877667,
"grad_norm": 2.357010841369629,
"learning_rate": 1.9775552389476865e-05,
"loss": 0.7818,
"step": 159
},
{
"epoch": 0.48553816974869607,
"grad_norm": 2.1862401962280273,
"learning_rate": 1.9768037755017687e-05,
"loss": 0.7879,
"step": 160
},
{
"epoch": 0.4885727833096254,
"grad_norm": 2.3130927085876465,
"learning_rate": 1.97604008616004e-05,
"loss": 0.7914,
"step": 161
},
{
"epoch": 0.49160739687055477,
"grad_norm": 2.0661509037017822,
"learning_rate": 1.9752641804810083e-05,
"loss": 0.8023,
"step": 162
},
{
"epoch": 0.4946420104314841,
"grad_norm": 2.2117955684661865,
"learning_rate": 1.9744760681760832e-05,
"loss": 0.7972,
"step": 163
},
{
"epoch": 0.49767662399241347,
"grad_norm": 2.603163242340088,
"learning_rate": 1.973675759109456e-05,
"loss": 0.7913,
"step": 164
},
{
"epoch": 0.5007112375533428,
"grad_norm": 2.2511062622070312,
"learning_rate": 1.9728632632979746e-05,
"loss": 0.7914,
"step": 165
},
{
"epoch": 0.5037458511142722,
"grad_norm": 2.375213861465454,
"learning_rate": 1.9720385909110197e-05,
"loss": 0.7928,
"step": 166
},
{
"epoch": 0.5067804646752015,
"grad_norm": 2.5348660945892334,
"learning_rate": 1.9712017522703764e-05,
"loss": 0.7894,
"step": 167
},
{
"epoch": 0.5098150782361308,
"grad_norm": 2.2094035148620605,
"learning_rate": 1.9703527578501052e-05,
"loss": 0.7813,
"step": 168
},
{
"epoch": 0.5128496917970602,
"grad_norm": 2.2283380031585693,
"learning_rate": 1.9694916182764113e-05,
"loss": 0.7877,
"step": 169
},
{
"epoch": 0.5158843053579896,
"grad_norm": 2.189119338989258,
"learning_rate": 1.9686183443275118e-05,
"loss": 0.7989,
"step": 170
},
{
"epoch": 0.518918918918919,
"grad_norm": 2.2636640071868896,
"learning_rate": 1.967732946933499e-05,
"loss": 0.8059,
"step": 171
},
{
"epoch": 0.5219535324798482,
"grad_norm": 2.243251085281372,
"learning_rate": 1.9668354371762066e-05,
"loss": 0.7904,
"step": 172
},
{
"epoch": 0.5249881460407776,
"grad_norm": 2.215536117553711,
"learning_rate": 1.9659258262890683e-05,
"loss": 0.7912,
"step": 173
},
{
"epoch": 0.528022759601707,
"grad_norm": 2.2998199462890625,
"learning_rate": 1.9650041256569792e-05,
"loss": 0.797,
"step": 174
},
{
"epoch": 0.5310573731626363,
"grad_norm": 2.3586950302124023,
"learning_rate": 1.9640703468161508e-05,
"loss": 0.7907,
"step": 175
},
{
"epoch": 0.5340919867235657,
"grad_norm": 2.257404088973999,
"learning_rate": 1.96312450145397e-05,
"loss": 0.7977,
"step": 176
},
{
"epoch": 0.537126600284495,
"grad_norm": 2.4808526039123535,
"learning_rate": 1.9621666014088495e-05,
"loss": 0.7929,
"step": 177
},
{
"epoch": 0.5401612138454244,
"grad_norm": 2.208704710006714,
"learning_rate": 1.9611966586700825e-05,
"loss": 0.7975,
"step": 178
},
{
"epoch": 0.5431958274063537,
"grad_norm": 2.4615161418914795,
"learning_rate": 1.9602146853776894e-05,
"loss": 0.7991,
"step": 179
},
{
"epoch": 0.5462304409672831,
"grad_norm": 2.4766757488250732,
"learning_rate": 1.9592206938222703e-05,
"loss": 0.7911,
"step": 180
},
{
"epoch": 0.5492650545282124,
"grad_norm": 2.2009432315826416,
"learning_rate": 1.9582146964448457e-05,
"loss": 0.788,
"step": 181
},
{
"epoch": 0.5522996680891418,
"grad_norm": 2.258129358291626,
"learning_rate": 1.9571967058367067e-05,
"loss": 0.7893,
"step": 182
},
{
"epoch": 0.5553342816500711,
"grad_norm": 2.1914985179901123,
"learning_rate": 1.956166734739251e-05,
"loss": 0.8057,
"step": 183
},
{
"epoch": 0.5583688952110005,
"grad_norm": 2.2815279960632324,
"learning_rate": 1.9551247960438298e-05,
"loss": 0.7823,
"step": 184
},
{
"epoch": 0.5614035087719298,
"grad_norm": 2.2393579483032227,
"learning_rate": 1.954070902791582e-05,
"loss": 0.7899,
"step": 185
},
{
"epoch": 0.5644381223328592,
"grad_norm": 2.56640625,
"learning_rate": 1.953005068173272e-05,
"loss": 0.7731,
"step": 186
},
{
"epoch": 0.5674727358937885,
"grad_norm": 2.3896234035491943,
"learning_rate": 1.9519273055291266e-05,
"loss": 0.7936,
"step": 187
},
{
"epoch": 0.5705073494547179,
"grad_norm": 2.21494722366333,
"learning_rate": 1.9508376283486653e-05,
"loss": 0.8121,
"step": 188
},
{
"epoch": 0.5735419630156472,
"grad_norm": 2.400538444519043,
"learning_rate": 1.949736050270532e-05,
"loss": 0.7938,
"step": 189
},
{
"epoch": 0.5765765765765766,
"grad_norm": 2.1337621212005615,
"learning_rate": 1.9486225850823265e-05,
"loss": 0.8049,
"step": 190
},
{
"epoch": 0.579611190137506,
"grad_norm": 2.314168930053711,
"learning_rate": 1.9474972467204298e-05,
"loss": 0.8109,
"step": 191
},
{
"epoch": 0.5826458036984353,
"grad_norm": 2.2364132404327393,
"learning_rate": 1.9463600492698297e-05,
"loss": 0.78,
"step": 192
},
{
"epoch": 0.5856804172593646,
"grad_norm": 2.204206705093384,
"learning_rate": 1.945211006963945e-05,
"loss": 0.8187,
"step": 193
},
{
"epoch": 0.588715030820294,
"grad_norm": 2.1865625381469727,
"learning_rate": 1.9440501341844484e-05,
"loss": 0.7844,
"step": 194
},
{
"epoch": 0.5917496443812233,
"grad_norm": 2.143092155456543,
"learning_rate": 1.9428774454610845e-05,
"loss": 0.7967,
"step": 195
},
{
"epoch": 0.5947842579421527,
"grad_norm": 2.3440985679626465,
"learning_rate": 1.9416929554714887e-05,
"loss": 0.8096,
"step": 196
},
{
"epoch": 0.597818871503082,
"grad_norm": 2.43977689743042,
"learning_rate": 1.9404966790410047e-05,
"loss": 0.795,
"step": 197
},
{
"epoch": 0.6008534850640114,
"grad_norm": 2.2740390300750732,
"learning_rate": 1.9392886311424975e-05,
"loss": 0.7937,
"step": 198
},
{
"epoch": 0.6038880986249408,
"grad_norm": 2.1775126457214355,
"learning_rate": 1.938068826896166e-05,
"loss": 0.8019,
"step": 199
},
{
"epoch": 0.6069227121858701,
"grad_norm": 2.1858487129211426,
"learning_rate": 1.9368372815693547e-05,
"loss": 0.7962,
"step": 200
},
{
"epoch": 0.6069227121858701,
"eval_loss": NaN,
"eval_runtime": 204.6421,
"eval_samples_per_second": 20.318,
"eval_steps_per_second": 5.082,
"step": 200
},
{
"epoch": 0.6099573257467994,
"grad_norm": 2.2958755493164062,
"learning_rate": 1.9355940105763622e-05,
"loss": 0.8003,
"step": 201
},
{
"epoch": 0.6129919393077288,
"grad_norm": 2.1330931186676025,
"learning_rate": 1.934339029478248e-05,
"loss": 0.7803,
"step": 202
},
{
"epoch": 0.6160265528686582,
"grad_norm": 2.3881402015686035,
"learning_rate": 1.9330723539826373e-05,
"loss": 0.8044,
"step": 203
},
{
"epoch": 0.6190611664295875,
"grad_norm": 2.2404513359069824,
"learning_rate": 1.9317939999435262e-05,
"loss": 0.8097,
"step": 204
},
{
"epoch": 0.6220957799905168,
"grad_norm": 2.194645404815674,
"learning_rate": 1.930503983361081e-05,
"loss": 0.8178,
"step": 205
},
{
"epoch": 0.6251303935514462,
"grad_norm": 2.509723424911499,
"learning_rate": 1.92920232038144e-05,
"loss": 0.7936,
"step": 206
},
{
"epoch": 0.6281650071123756,
"grad_norm": 2.2457869052886963,
"learning_rate": 1.9278890272965097e-05,
"loss": 0.8162,
"step": 207
},
{
"epoch": 0.6311996206733049,
"grad_norm": 2.2989683151245117,
"learning_rate": 1.9265641205437612e-05,
"loss": 0.8012,
"step": 208
},
{
"epoch": 0.6342342342342342,
"grad_norm": 2.3188092708587646,
"learning_rate": 1.925227616706026e-05,
"loss": 0.8094,
"step": 209
},
{
"epoch": 0.6372688477951636,
"grad_norm": 2.2978076934814453,
"learning_rate": 1.9238795325112867e-05,
"loss": 0.7811,
"step": 210
},
{
"epoch": 0.640303461356093,
"grad_norm": 2.331542730331421,
"learning_rate": 1.9225198848324687e-05,
"loss": 0.7946,
"step": 211
},
{
"epoch": 0.6433380749170223,
"grad_norm": 2.1989738941192627,
"learning_rate": 1.921148690687228e-05,
"loss": 0.7983,
"step": 212
},
{
"epoch": 0.6463726884779516,
"grad_norm": 2.4261419773101807,
"learning_rate": 1.9197659672377388e-05,
"loss": 0.8134,
"step": 213
},
{
"epoch": 0.649407302038881,
"grad_norm": 2.3790082931518555,
"learning_rate": 1.918371731790479e-05,
"loss": 0.79,
"step": 214
},
{
"epoch": 0.6524419155998104,
"grad_norm": 2.2966949939727783,
"learning_rate": 1.9169660017960135e-05,
"loss": 0.8029,
"step": 215
},
{
"epoch": 0.6554765291607397,
"grad_norm": 2.5911426544189453,
"learning_rate": 1.915548794848775e-05,
"loss": 0.8118,
"step": 216
},
{
"epoch": 0.658511142721669,
"grad_norm": 2.3282856941223145,
"learning_rate": 1.9141201286868435e-05,
"loss": 0.8092,
"step": 217
},
{
"epoch": 0.6615457562825984,
"grad_norm": 2.415398359298706,
"learning_rate": 1.9126800211917277e-05,
"loss": 0.8156,
"step": 218
},
{
"epoch": 0.6645803698435278,
"grad_norm": 2.2823410034179688,
"learning_rate": 1.911228490388136e-05,
"loss": 0.8004,
"step": 219
},
{
"epoch": 0.6676149834044571,
"grad_norm": 2.2104527950286865,
"learning_rate": 1.9097655544437544e-05,
"loss": 0.8023,
"step": 220
},
{
"epoch": 0.6706495969653864,
"grad_norm": 2.3354063034057617,
"learning_rate": 1.908291231669019e-05,
"loss": 0.8117,
"step": 221
},
{
"epoch": 0.6736842105263158,
"grad_norm": 2.230656147003174,
"learning_rate": 1.906805540516885e-05,
"loss": 0.7797,
"step": 222
},
{
"epoch": 0.6767188240872452,
"grad_norm": 2.464111328125,
"learning_rate": 1.905308499582597e-05,
"loss": 0.7929,
"step": 223
},
{
"epoch": 0.6797534376481745,
"grad_norm": 2.188788652420044,
"learning_rate": 1.903800127603456e-05,
"loss": 0.7971,
"step": 224
},
{
"epoch": 0.6827880512091038,
"grad_norm": 2.202427387237549,
"learning_rate": 1.9022804434585854e-05,
"loss": 0.8026,
"step": 225
},
{
"epoch": 0.6858226647700332,
"grad_norm": 2.2621190547943115,
"learning_rate": 1.9007494661686937e-05,
"loss": 0.8112,
"step": 226
},
{
"epoch": 0.6888572783309626,
"grad_norm": 2.333603620529175,
"learning_rate": 1.8992072148958368e-05,
"loss": 0.7937,
"step": 227
},
{
"epoch": 0.6918918918918919,
"grad_norm": 2.626451253890991,
"learning_rate": 1.8976537089431793e-05,
"loss": 0.8005,
"step": 228
},
{
"epoch": 0.6949265054528212,
"grad_norm": 2.4107227325439453,
"learning_rate": 1.8960889677547506e-05,
"loss": 0.7813,
"step": 229
},
{
"epoch": 0.6979611190137506,
"grad_norm": 2.7147607803344727,
"learning_rate": 1.8945130109152035e-05,
"loss": 0.8036,
"step": 230
},
{
"epoch": 0.70099573257468,
"grad_norm": 2.281703233718872,
"learning_rate": 1.8929258581495688e-05,
"loss": 0.7946,
"step": 231
},
{
"epoch": 0.7040303461356093,
"grad_norm": 3.016942262649536,
"learning_rate": 1.891327529323007e-05,
"loss": 0.7786,
"step": 232
},
{
"epoch": 0.7070649596965386,
"grad_norm": 2.5729317665100098,
"learning_rate": 1.8897180444405615e-05,
"loss": 0.8141,
"step": 233
},
{
"epoch": 0.710099573257468,
"grad_norm": 2.75722336769104,
"learning_rate": 1.888097423646907e-05,
"loss": 0.8079,
"step": 234
},
{
"epoch": 0.7131341868183974,
"grad_norm": 2.843980073928833,
"learning_rate": 1.8864656872260985e-05,
"loss": 0.795,
"step": 235
},
{
"epoch": 0.7161688003793267,
"grad_norm": 3.3999879360198975,
"learning_rate": 1.884822855601316e-05,
"loss": 0.8086,
"step": 236
},
{
"epoch": 0.719203413940256,
"grad_norm": 3.1997487545013428,
"learning_rate": 1.8831689493346095e-05,
"loss": 0.7919,
"step": 237
},
{
"epoch": 0.7222380275011854,
"grad_norm": 2.4851367473602295,
"learning_rate": 1.881503989126642e-05,
"loss": 0.7914,
"step": 238
},
{
"epoch": 0.7252726410621148,
"grad_norm": 2.738428831100464,
"learning_rate": 1.8798279958164295e-05,
"loss": 0.7929,
"step": 239
},
{
"epoch": 0.7283072546230441,
"grad_norm": 2.1467783451080322,
"learning_rate": 1.8781409903810823e-05,
"loss": 0.815,
"step": 240
},
{
"epoch": 0.7313418681839734,
"grad_norm": 16.773521423339844,
"learning_rate": 1.8764429939355394e-05,
"loss": 0.8014,
"step": 241
},
{
"epoch": 0.7343764817449028,
"grad_norm": 3.171633720397949,
"learning_rate": 1.874734027732306e-05,
"loss": 0.805,
"step": 242
},
{
"epoch": 0.7374110953058322,
"grad_norm": 2.249514102935791,
"learning_rate": 1.8730141131611882e-05,
"loss": 0.8109,
"step": 243
},
{
"epoch": 0.7404457088667615,
"grad_norm": 2.816488742828369,
"learning_rate": 1.8712832717490238e-05,
"loss": 0.8044,
"step": 244
},
{
"epoch": 0.7434803224276908,
"grad_norm": 2.558295488357544,
"learning_rate": 1.8695415251594123e-05,
"loss": 0.8269,
"step": 245
},
{
"epoch": 0.7465149359886202,
"grad_norm": 2.5777034759521484,
"learning_rate": 1.8677888951924473e-05,
"loss": 0.7971,
"step": 246
},
{
"epoch": 0.7495495495495496,
"grad_norm": 2.394287586212158,
"learning_rate": 1.866025403784439e-05,
"loss": 0.8092,
"step": 247
},
{
"epoch": 0.7525841631104789,
"grad_norm": 2.2798614501953125,
"learning_rate": 1.864251073007642e-05,
"loss": 0.7964,
"step": 248
},
{
"epoch": 0.7556187766714082,
"grad_norm": 2.3587262630462646,
"learning_rate": 1.8624659250699807e-05,
"loss": 0.7928,
"step": 249
},
{
"epoch": 0.7586533902323376,
"grad_norm": 2.189763307571411,
"learning_rate": 1.8606699823147675e-05,
"loss": 0.7812,
"step": 250
},
{
"epoch": 0.7586533902323376,
"eval_loss": NaN,
"eval_runtime": 204.0596,
"eval_samples_per_second": 20.376,
"eval_steps_per_second": 5.097,
"step": 250
},
{
"epoch": 0.761688003793267,
"grad_norm": 2.433803081512451,
"learning_rate": 1.8588632672204264e-05,
"loss": 0.8111,
"step": 251
},
{
"epoch": 0.7647226173541963,
"grad_norm": 2.28267765045166,
"learning_rate": 1.8570458024002094e-05,
"loss": 0.8001,
"step": 252
},
{
"epoch": 0.7677572309151256,
"grad_norm": 2.3339545726776123,
"learning_rate": 1.8552176106019156e-05,
"loss": 0.8158,
"step": 253
},
{
"epoch": 0.770791844476055,
"grad_norm": 2.284759998321533,
"learning_rate": 1.8533787147076046e-05,
"loss": 0.7852,
"step": 254
},
{
"epoch": 0.7738264580369844,
"grad_norm": 2.35969614982605,
"learning_rate": 1.8515291377333114e-05,
"loss": 0.7909,
"step": 255
},
{
"epoch": 0.7768610715979137,
"grad_norm": 2.4525341987609863,
"learning_rate": 1.8496689028287572e-05,
"loss": 0.817,
"step": 256
},
{
"epoch": 0.779895685158843,
"grad_norm": 2.7126755714416504,
"learning_rate": 1.847798033277061e-05,
"loss": 0.797,
"step": 257
},
{
"epoch": 0.7829302987197724,
"grad_norm": 2.2292561531066895,
"learning_rate": 1.8459165524944463e-05,
"loss": 0.8044,
"step": 258
},
{
"epoch": 0.7859649122807018,
"grad_norm": 2.447347640991211,
"learning_rate": 1.8440244840299507e-05,
"loss": 0.7979,
"step": 259
},
{
"epoch": 0.7889995258416311,
"grad_norm": 2.2336087226867676,
"learning_rate": 1.842121851565128e-05,
"loss": 0.8036,
"step": 260
},
{
"epoch": 0.7920341394025604,
"grad_norm": 2.3720502853393555,
"learning_rate": 1.8402086789137547e-05,
"loss": 0.7979,
"step": 261
},
{
"epoch": 0.7950687529634898,
"grad_norm": 2.1780807971954346,
"learning_rate": 1.8382849900215297e-05,
"loss": 0.7876,
"step": 262
},
{
"epoch": 0.7981033665244192,
"grad_norm": 2.3325858116149902,
"learning_rate": 1.8363508089657763e-05,
"loss": 0.7997,
"step": 263
},
{
"epoch": 0.8011379800853485,
"grad_norm": 2.3341164588928223,
"learning_rate": 1.8344061599551397e-05,
"loss": 0.7844,
"step": 264
},
{
"epoch": 0.8041725936462779,
"grad_norm": 2.4678280353546143,
"learning_rate": 1.8324510673292844e-05,
"loss": 0.7946,
"step": 265
},
{
"epoch": 0.8072072072072072,
"grad_norm": 2.424893856048584,
"learning_rate": 1.8304855555585893e-05,
"loss": 0.7916,
"step": 266
},
{
"epoch": 0.8102418207681366,
"grad_norm": 2.5433976650238037,
"learning_rate": 1.8285096492438424e-05,
"loss": 0.7983,
"step": 267
},
{
"epoch": 0.8132764343290659,
"grad_norm": 2.499178647994995,
"learning_rate": 1.826523373115931e-05,
"loss": 0.7944,
"step": 268
},
{
"epoch": 0.8163110478899953,
"grad_norm": 2.2453994750976562,
"learning_rate": 1.8245267520355348e-05,
"loss": 0.8148,
"step": 269
},
{
"epoch": 0.8193456614509246,
"grad_norm": 2.350146770477295,
"learning_rate": 1.8225198109928116e-05,
"loss": 0.7986,
"step": 270
},
{
"epoch": 0.822380275011854,
"grad_norm": 2.2048559188842773,
"learning_rate": 1.8205025751070878e-05,
"loss": 0.8093,
"step": 271
},
{
"epoch": 0.8254148885727833,
"grad_norm": 2.2730185985565186,
"learning_rate": 1.8184750696265408e-05,
"loss": 0.7787,
"step": 272
},
{
"epoch": 0.8284495021337127,
"grad_norm": 2.423301935195923,
"learning_rate": 1.8164373199278858e-05,
"loss": 0.823,
"step": 273
},
{
"epoch": 0.831484115694642,
"grad_norm": 2.309649705886841,
"learning_rate": 1.8143893515160565e-05,
"loss": 0.7901,
"step": 274
},
{
"epoch": 0.8345187292555714,
"grad_norm": 2.249284267425537,
"learning_rate": 1.812331190023886e-05,
"loss": 0.8095,
"step": 275
},
{
"epoch": 0.8375533428165007,
"grad_norm": 2.2063703536987305,
"learning_rate": 1.8102628612117868e-05,
"loss": 0.8008,
"step": 276
},
{
"epoch": 0.8405879563774301,
"grad_norm": 2.2518839836120605,
"learning_rate": 1.8081843909674277e-05,
"loss": 0.8051,
"step": 277
},
{
"epoch": 0.8436225699383594,
"grad_norm": 2.226356267929077,
"learning_rate": 1.8060958053054095e-05,
"loss": 0.8036,
"step": 278
},
{
"epoch": 0.8466571834992888,
"grad_norm": 2.186485767364502,
"learning_rate": 1.8039971303669407e-05,
"loss": 0.8025,
"step": 279
},
{
"epoch": 0.8496917970602181,
"grad_norm": 2.4235646724700928,
"learning_rate": 1.8018883924195085e-05,
"loss": 0.7799,
"step": 280
},
{
"epoch": 0.8527264106211475,
"grad_norm": 2.2470943927764893,
"learning_rate": 1.799769617856552e-05,
"loss": 0.8025,
"step": 281
},
{
"epoch": 0.8557610241820768,
"grad_norm": 2.2514889240264893,
"learning_rate": 1.79764083319713e-05,
"loss": 0.7978,
"step": 282
},
{
"epoch": 0.8587956377430062,
"grad_norm": 2.220952033996582,
"learning_rate": 1.79550206508559e-05,
"loss": 0.7978,
"step": 283
},
{
"epoch": 0.8618302513039355,
"grad_norm": 2.4699270725250244,
"learning_rate": 1.7933533402912354e-05,
"loss": 0.7835,
"step": 284
},
{
"epoch": 0.8648648648648649,
"grad_norm": 2.3011207580566406,
"learning_rate": 1.7911946857079886e-05,
"loss": 0.8008,
"step": 285
},
{
"epoch": 0.8678994784257942,
"grad_norm": 2.2239327430725098,
"learning_rate": 1.7890261283540563e-05,
"loss": 0.7904,
"step": 286
},
{
"epoch": 0.8709340919867236,
"grad_norm": 2.077845573425293,
"learning_rate": 1.78684769537159e-05,
"loss": 0.7962,
"step": 287
},
{
"epoch": 0.8739687055476529,
"grad_norm": 2.2492687702178955,
"learning_rate": 1.7846594140263475e-05,
"loss": 0.8076,
"step": 288
},
{
"epoch": 0.8770033191085823,
"grad_norm": 2.1000773906707764,
"learning_rate": 1.78246131170735e-05,
"loss": 0.8107,
"step": 289
},
{
"epoch": 0.8800379326695116,
"grad_norm": 2.2440242767333984,
"learning_rate": 1.7802534159265407e-05,
"loss": 0.784,
"step": 290
},
{
"epoch": 0.883072546230441,
"grad_norm": 2.1370065212249756,
"learning_rate": 1.7780357543184396e-05,
"loss": 0.7926,
"step": 291
},
{
"epoch": 0.8861071597913703,
"grad_norm": 2.2182703018188477,
"learning_rate": 1.775808354639799e-05,
"loss": 0.8079,
"step": 292
},
{
"epoch": 0.8891417733522997,
"grad_norm": 2.236370325088501,
"learning_rate": 1.773571244769254e-05,
"loss": 0.7851,
"step": 293
},
{
"epoch": 0.892176386913229,
"grad_norm": 2.216042995452881,
"learning_rate": 1.771324452706975e-05,
"loss": 0.798,
"step": 294
},
{
"epoch": 0.8952110004741584,
"grad_norm": 2.2804715633392334,
"learning_rate": 1.769068006574317e-05,
"loss": 0.7916,
"step": 295
},
{
"epoch": 0.8982456140350877,
"grad_norm": 2.188271999359131,
"learning_rate": 1.7668019346134674e-05,
"loss": 0.7993,
"step": 296
},
{
"epoch": 0.9012802275960171,
"grad_norm": 2.372596025466919,
"learning_rate": 1.7645262651870926e-05,
"loss": 0.816,
"step": 297
},
{
"epoch": 0.9043148411569464,
"grad_norm": 2.174302339553833,
"learning_rate": 1.7622410267779834e-05,
"loss": 0.8247,
"step": 298
},
{
"epoch": 0.9073494547178758,
"grad_norm": 2.0944302082061768,
"learning_rate": 1.7599462479886976e-05,
"loss": 0.7979,
"step": 299
},
{
"epoch": 0.9103840682788051,
"grad_norm": 2.118502140045166,
"learning_rate": 1.7576419575412028e-05,
"loss": 0.8007,
"step": 300
},
{
"epoch": 0.9103840682788051,
"eval_loss": NaN,
"eval_runtime": 203.6703,
"eval_samples_per_second": 20.415,
"eval_steps_per_second": 5.106,
"step": 300
},
{
"epoch": 0.9134186818397345,
"grad_norm": 2.2053334712982178,
"learning_rate": 1.755328184276517e-05,
"loss": 0.7913,
"step": 301
},
{
"epoch": 0.9164532954006638,
"grad_norm": 2.243788957595825,
"learning_rate": 1.7530049571543464e-05,
"loss": 0.801,
"step": 302
},
{
"epoch": 0.9194879089615932,
"grad_norm": 2.363306760787964,
"learning_rate": 1.7506723052527243e-05,
"loss": 0.8278,
"step": 303
},
{
"epoch": 0.9225225225225225,
"grad_norm": 2.2566967010498047,
"learning_rate": 1.7483302577676475e-05,
"loss": 0.7929,
"step": 304
},
{
"epoch": 0.9255571360834519,
"grad_norm": 2.1566100120544434,
"learning_rate": 1.7459788440127083e-05,
"loss": 0.7953,
"step": 305
},
{
"epoch": 0.9285917496443812,
"grad_norm": 2.279130697250366,
"learning_rate": 1.7436180934187307e-05,
"loss": 0.8125,
"step": 306
},
{
"epoch": 0.9316263632053106,
"grad_norm": 2.172891616821289,
"learning_rate": 1.7412480355334006e-05,
"loss": 0.8007,
"step": 307
},
{
"epoch": 0.9346609767662399,
"grad_norm": 2.2873098850250244,
"learning_rate": 1.738868700020895e-05,
"loss": 0.779,
"step": 308
},
{
"epoch": 0.9376955903271693,
"grad_norm": 2.3155357837677,
"learning_rate": 1.7364801166615124e-05,
"loss": 0.8025,
"step": 309
},
{
"epoch": 0.9407302038880986,
"grad_norm": 2.20151686668396,
"learning_rate": 1.7340823153513003e-05,
"loss": 0.8173,
"step": 310
},
{
"epoch": 0.943764817449028,
"grad_norm": 2.2033958435058594,
"learning_rate": 1.7316753261016782e-05,
"loss": 0.8095,
"step": 311
},
{
"epoch": 0.9467994310099573,
"grad_norm": 2.2805070877075195,
"learning_rate": 1.7292591790390668e-05,
"loss": 0.8139,
"step": 312
},
{
"epoch": 0.9498340445708867,
"grad_norm": 2.335238456726074,
"learning_rate": 1.7268339044045044e-05,
"loss": 0.7898,
"step": 313
},
{
"epoch": 0.952868658131816,
"grad_norm": 2.1889333724975586,
"learning_rate": 1.7243995325532755e-05,
"loss": 0.8002,
"step": 314
},
{
"epoch": 0.9559032716927454,
"grad_norm": 2.5007987022399902,
"learning_rate": 1.7219560939545246e-05,
"loss": 0.7892,
"step": 315
},
{
"epoch": 0.9589378852536747,
"grad_norm": 2.1342813968658447,
"learning_rate": 1.7195036191908798e-05,
"loss": 0.8028,
"step": 316
},
{
"epoch": 0.9619724988146041,
"grad_norm": 2.199307918548584,
"learning_rate": 1.7170421389580666e-05,
"loss": 0.793,
"step": 317
},
{
"epoch": 0.9650071123755334,
"grad_norm": 2.1955904960632324,
"learning_rate": 1.7145716840645253e-05,
"loss": 0.8085,
"step": 318
},
{
"epoch": 0.9680417259364628,
"grad_norm": 2.41741943359375,
"learning_rate": 1.712092285431026e-05,
"loss": 0.7964,
"step": 319
},
{
"epoch": 0.9710763394973921,
"grad_norm": 2.3595402240753174,
"learning_rate": 1.7096039740902782e-05,
"loss": 0.7999,
"step": 320
},
{
"epoch": 0.9741109530583215,
"grad_norm": 2.15049147605896,
"learning_rate": 1.7071067811865477e-05,
"loss": 0.7813,
"step": 321
},
{
"epoch": 0.9771455666192508,
"grad_norm": 2.240618944168091,
"learning_rate": 1.7046007379752624e-05,
"loss": 0.8038,
"step": 322
},
{
"epoch": 0.9801801801801802,
"grad_norm": 2.418973922729492,
"learning_rate": 1.702085875822623e-05,
"loss": 0.7672,
"step": 323
},
{
"epoch": 0.9832147937411095,
"grad_norm": 2.2924294471740723,
"learning_rate": 1.6995622262052093e-05,
"loss": 0.8013,
"step": 324
},
{
"epoch": 0.9862494073020389,
"grad_norm": 2.2135136127471924,
"learning_rate": 1.6970298207095887e-05,
"loss": 0.8112,
"step": 325
},
{
"epoch": 0.9892840208629682,
"grad_norm": 2.2720751762390137,
"learning_rate": 1.6944886910319173e-05,
"loss": 0.7896,
"step": 326
},
{
"epoch": 0.9923186344238976,
"grad_norm": 2.172165632247925,
"learning_rate": 1.6919388689775463e-05,
"loss": 0.8084,
"step": 327
},
{
"epoch": 0.9953532479848269,
"grad_norm": 2.1608591079711914,
"learning_rate": 1.6893803864606224e-05,
"loss": 0.7914,
"step": 328
},
{
"epoch": 0.9983878615457563,
"grad_norm": 2.153231143951416,
"learning_rate": 1.6868132755036875e-05,
"loss": 0.803,
"step": 329
},
{
"epoch": 1.0014224751066856,
"grad_norm": 2.242035150527954,
"learning_rate": 1.6842375682372803e-05,
"loss": 0.7964,
"step": 330
},
{
"epoch": 1.004457088667615,
"grad_norm": 2.490118980407715,
"learning_rate": 1.681653296899533e-05,
"loss": 0.7645,
"step": 331
},
{
"epoch": 1.0074917022285443,
"grad_norm": 2.258561372756958,
"learning_rate": 1.6790604938357664e-05,
"loss": 0.7579,
"step": 332
},
{
"epoch": 1.0105263157894737,
"grad_norm": 2.3750267028808594,
"learning_rate": 1.676459191498087e-05,
"loss": 0.7849,
"step": 333
},
{
"epoch": 1.013560929350403,
"grad_norm": 2.5088164806365967,
"learning_rate": 1.6738494224449802e-05,
"loss": 0.7814,
"step": 334
},
{
"epoch": 1.0165955429113325,
"grad_norm": 2.4220824241638184,
"learning_rate": 1.6712312193409032e-05,
"loss": 0.7774,
"step": 335
},
{
"epoch": 1.0196301564722616,
"grad_norm": 2.2030773162841797,
"learning_rate": 1.6686046149558736e-05,
"loss": 0.7495,
"step": 336
},
{
"epoch": 1.022664770033191,
"grad_norm": 2.336583375930786,
"learning_rate": 1.6659696421650645e-05,
"loss": 0.7508,
"step": 337
},
{
"epoch": 1.0256993835941204,
"grad_norm": 2.3112359046936035,
"learning_rate": 1.6633263339483867e-05,
"loss": 0.7525,
"step": 338
},
{
"epoch": 1.0287339971550498,
"grad_norm": 2.165022611618042,
"learning_rate": 1.6606747233900816e-05,
"loss": 0.787,
"step": 339
},
{
"epoch": 1.0317686107159791,
"grad_norm": 2.3545596599578857,
"learning_rate": 1.658014843678303e-05,
"loss": 0.7665,
"step": 340
},
{
"epoch": 1.0348032242769085,
"grad_norm": 2.231351137161255,
"learning_rate": 1.655346728104704e-05,
"loss": 0.7726,
"step": 341
},
{
"epoch": 1.037837837837838,
"grad_norm": 2.292428970336914,
"learning_rate": 1.652670410064019e-05,
"loss": 0.7722,
"step": 342
},
{
"epoch": 1.0408724513987673,
"grad_norm": 2.2555713653564453,
"learning_rate": 1.6499859230536468e-05,
"loss": 0.755,
"step": 343
},
{
"epoch": 1.0439070649596967,
"grad_norm": 2.5129449367523193,
"learning_rate": 1.647293300673231e-05,
"loss": 0.7736,
"step": 344
},
{
"epoch": 1.0469416785206258,
"grad_norm": 2.269122362136841,
"learning_rate": 1.6445925766242392e-05,
"loss": 0.7898,
"step": 345
},
{
"epoch": 1.0499762920815552,
"grad_norm": 2.21991229057312,
"learning_rate": 1.641883784709541e-05,
"loss": 0.7767,
"step": 346
},
{
"epoch": 1.0530109056424846,
"grad_norm": 2.324253797531128,
"learning_rate": 1.639166958832985e-05,
"loss": 0.7728,
"step": 347
},
{
"epoch": 1.056045519203414,
"grad_norm": 2.3205628395080566,
"learning_rate": 1.6364421329989758e-05,
"loss": 0.7845,
"step": 348
},
{
"epoch": 1.0590801327643433,
"grad_norm": 2.361678123474121,
"learning_rate": 1.6337093413120463e-05,
"loss": 0.7455,
"step": 349
},
{
"epoch": 1.0621147463252727,
"grad_norm": 2.3375606536865234,
"learning_rate": 1.6309686179764317e-05,
"loss": 0.7754,
"step": 350
},
{
"epoch": 1.0621147463252727,
"eval_loss": NaN,
"eval_runtime": 203.6306,
"eval_samples_per_second": 20.419,
"eval_steps_per_second": 5.107,
"step": 350
},
{
"epoch": 1.065149359886202,
"grad_norm": 2.3522422313690186,
"learning_rate": 1.6282199972956425e-05,
"loss": 0.7759,
"step": 351
},
{
"epoch": 1.0681839734471312,
"grad_norm": 2.4227213859558105,
"learning_rate": 1.6254635136720328e-05,
"loss": 0.7772,
"step": 352
},
{
"epoch": 1.0712185870080606,
"grad_norm": 2.235722303390503,
"learning_rate": 1.6226992016063726e-05,
"loss": 0.7694,
"step": 353
},
{
"epoch": 1.07425320056899,
"grad_norm": 2.2417314052581787,
"learning_rate": 1.6199270956974128e-05,
"loss": 0.7628,
"step": 354
},
{
"epoch": 1.0772878141299194,
"grad_norm": 2.348954200744629,
"learning_rate": 1.6171472306414554e-05,
"loss": 0.7656,
"step": 355
},
{
"epoch": 1.0803224276908487,
"grad_norm": 2.346963882446289,
"learning_rate": 1.614359641231916e-05,
"loss": 0.7839,
"step": 356
},
{
"epoch": 1.0833570412517781,
"grad_norm": 2.2969138622283936,
"learning_rate": 1.6115643623588915e-05,
"loss": 0.7728,
"step": 357
},
{
"epoch": 1.0863916548127075,
"grad_norm": 2.3338327407836914,
"learning_rate": 1.608761429008721e-05,
"loss": 0.7902,
"step": 358
},
{
"epoch": 1.0894262683736369,
"grad_norm": 2.2462401390075684,
"learning_rate": 1.6059508762635482e-05,
"loss": 0.7906,
"step": 359
},
{
"epoch": 1.0924608819345663,
"grad_norm": 2.2091758251190186,
"learning_rate": 1.6031327393008848e-05,
"loss": 0.7587,
"step": 360
},
{
"epoch": 1.0954954954954954,
"grad_norm": 2.2392489910125732,
"learning_rate": 1.6003070533931657e-05,
"loss": 0.7598,
"step": 361
},
{
"epoch": 1.0985301090564248,
"grad_norm": 2.193833112716675,
"learning_rate": 1.5974738539073125e-05,
"loss": 0.7622,
"step": 362
},
{
"epoch": 1.1015647226173542,
"grad_norm": 2.187610149383545,
"learning_rate": 1.594633176304287e-05,
"loss": 0.7796,
"step": 363
},
{
"epoch": 1.1045993361782835,
"grad_norm": 2.2753069400787354,
"learning_rate": 1.5917850561386487e-05,
"loss": 0.7783,
"step": 364
},
{
"epoch": 1.107633949739213,
"grad_norm": 2.2835614681243896,
"learning_rate": 1.588929529058111e-05,
"loss": 0.7801,
"step": 365
},
{
"epoch": 1.1106685633001423,
"grad_norm": 2.46498441696167,
"learning_rate": 1.5860666308030933e-05,
"loss": 0.7683,
"step": 366
},
{
"epoch": 1.1137031768610717,
"grad_norm": 2.361351490020752,
"learning_rate": 1.5831963972062734e-05,
"loss": 0.783,
"step": 367
},
{
"epoch": 1.1167377904220008,
"grad_norm": 2.2396347522735596,
"learning_rate": 1.5803188641921417e-05,
"loss": 0.7563,
"step": 368
},
{
"epoch": 1.1197724039829302,
"grad_norm": 2.2810609340667725,
"learning_rate": 1.5774340677765483e-05,
"loss": 0.7865,
"step": 369
},
{
"epoch": 1.1228070175438596,
"grad_norm": 2.147937297821045,
"learning_rate": 1.5745420440662543e-05,
"loss": 0.7684,
"step": 370
},
{
"epoch": 1.125841631104789,
"grad_norm": 2.355337142944336,
"learning_rate": 1.5716428292584788e-05,
"loss": 0.7693,
"step": 371
},
{
"epoch": 1.1288762446657183,
"grad_norm": 2.4234957695007324,
"learning_rate": 1.568736459640447e-05,
"loss": 0.7567,
"step": 372
},
{
"epoch": 1.1319108582266477,
"grad_norm": 2.2321126461029053,
"learning_rate": 1.5658229715889345e-05,
"loss": 0.7984,
"step": 373
},
{
"epoch": 1.134945471787577,
"grad_norm": 2.2270772457122803,
"learning_rate": 1.5629024015698137e-05,
"loss": 0.7868,
"step": 374
},
{
"epoch": 1.1379800853485065,
"grad_norm": 2.4906022548675537,
"learning_rate": 1.5599747861375957e-05,
"loss": 0.7761,
"step": 375
},
{
"epoch": 1.1410146989094359,
"grad_norm": 2.4099533557891846,
"learning_rate": 1.5570401619349737e-05,
"loss": 0.7727,
"step": 376
},
{
"epoch": 1.144049312470365,
"grad_norm": 2.167451858520508,
"learning_rate": 1.5540985656923648e-05,
"loss": 0.7449,
"step": 377
},
{
"epoch": 1.1470839260312944,
"grad_norm": 2.4137990474700928,
"learning_rate": 1.551150034227449e-05,
"loss": 0.7836,
"step": 378
},
{
"epoch": 1.1501185395922238,
"grad_norm": 2.0170676708221436,
"learning_rate": 1.54819460444471e-05,
"loss": 0.771,
"step": 379
},
{
"epoch": 1.1531531531531531,
"grad_norm": 2.3556909561157227,
"learning_rate": 1.5452323133349712e-05,
"loss": 0.769,
"step": 380
},
{
"epoch": 1.1561877667140825,
"grad_norm": 2.1490001678466797,
"learning_rate": 1.5422631979749354e-05,
"loss": 0.7467,
"step": 381
},
{
"epoch": 1.159222380275012,
"grad_norm": 2.240185260772705,
"learning_rate": 1.5392872955267176e-05,
"loss": 0.7677,
"step": 382
},
{
"epoch": 1.1622569938359413,
"grad_norm": 2.1402430534362793,
"learning_rate": 1.5363046432373824e-05,
"loss": 0.7706,
"step": 383
},
{
"epoch": 1.1652916073968704,
"grad_norm": 2.043536424636841,
"learning_rate": 1.5333152784384777e-05,
"loss": 0.7679,
"step": 384
},
{
"epoch": 1.1683262209577998,
"grad_norm": 2.146432399749756,
"learning_rate": 1.5303192385455652e-05,
"loss": 0.7746,
"step": 385
},
{
"epoch": 1.1713608345187292,
"grad_norm": 2.216525077819824,
"learning_rate": 1.5273165610577543e-05,
"loss": 0.7735,
"step": 386
},
{
"epoch": 1.1743954480796586,
"grad_norm": 2.134411334991455,
"learning_rate": 1.5243072835572319e-05,
"loss": 0.771,
"step": 387
},
{
"epoch": 1.177430061640588,
"grad_norm": 2.210275173187256,
"learning_rate": 1.5212914437087921e-05,
"loss": 0.7665,
"step": 388
},
{
"epoch": 1.1804646752015173,
"grad_norm": 2.2126176357269287,
"learning_rate": 1.5182690792593659e-05,
"loss": 0.7658,
"step": 389
},
{
"epoch": 1.1834992887624467,
"grad_norm": 1.9885146617889404,
"learning_rate": 1.5152402280375454e-05,
"loss": 0.7509,
"step": 390
},
{
"epoch": 1.186533902323376,
"grad_norm": 2.2027952671051025,
"learning_rate": 1.5122049279531143e-05,
"loss": 0.7811,
"step": 391
},
{
"epoch": 1.1895685158843055,
"grad_norm": 2.1936960220336914,
"learning_rate": 1.509163216996572e-05,
"loss": 0.7785,
"step": 392
},
{
"epoch": 1.1926031294452346,
"grad_norm": 2.166518449783325,
"learning_rate": 1.5061151332386565e-05,
"loss": 0.7775,
"step": 393
},
{
"epoch": 1.195637743006164,
"grad_norm": 2.0909955501556396,
"learning_rate": 1.5030607148298697e-05,
"loss": 0.7783,
"step": 394
},
{
"epoch": 1.1986723565670934,
"grad_norm": 2.287322521209717,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.7678,
"step": 395
},
{
"epoch": 1.2017069701280227,
"grad_norm": 2.265106678009033,
"learning_rate": 1.4969330270576428e-05,
"loss": 0.7772,
"step": 396
},
{
"epoch": 1.2047415836889521,
"grad_norm": 2.2475264072418213,
"learning_rate": 1.4938598343897215e-05,
"loss": 0.7509,
"step": 397
},
{
"epoch": 1.2077761972498815,
"grad_norm": 2.3240981101989746,
"learning_rate": 1.4907804604610064e-05,
"loss": 0.7849,
"step": 398
},
{
"epoch": 1.2108108108108109,
"grad_norm": 2.148869276046753,
"learning_rate": 1.4876949438136348e-05,
"loss": 0.7781,
"step": 399
},
{
"epoch": 1.2138454243717403,
"grad_norm": 2.318875789642334,
"learning_rate": 1.484603323066627e-05,
"loss": 0.7987,
"step": 400
},
{
"epoch": 1.2138454243717403,
"eval_loss": NaN,
"eval_runtime": 204.8025,
"eval_samples_per_second": 20.302,
"eval_steps_per_second": 5.078,
"step": 400
},
{
"epoch": 1.2168800379326696,
"grad_norm": 2.1469035148620605,
"learning_rate": 1.4815056369154039e-05,
"loss": 0.7924,
"step": 401
},
{
"epoch": 1.2199146514935988,
"grad_norm": 2.3983654975891113,
"learning_rate": 1.4784019241313025e-05,
"loss": 0.7431,
"step": 402
},
{
"epoch": 1.2229492650545282,
"grad_norm": 2.1171765327453613,
"learning_rate": 1.47529222356109e-05,
"loss": 0.7583,
"step": 403
},
{
"epoch": 1.2259838786154575,
"grad_norm": 2.3186557292938232,
"learning_rate": 1.4721765741264786e-05,
"loss": 0.7545,
"step": 404
},
{
"epoch": 1.229018492176387,
"grad_norm": 2.308945417404175,
"learning_rate": 1.4690550148236371e-05,
"loss": 0.7752,
"step": 405
},
{
"epoch": 1.2320531057373163,
"grad_norm": 2.141418933868408,
"learning_rate": 1.4659275847227044e-05,
"loss": 0.7501,
"step": 406
},
{
"epoch": 1.2350877192982457,
"grad_norm": 2.1447696685791016,
"learning_rate": 1.4627943229672992e-05,
"loss": 0.7446,
"step": 407
},
{
"epoch": 1.238122332859175,
"grad_norm": 2.062683582305908,
"learning_rate": 1.4596552687740304e-05,
"loss": 0.7729,
"step": 408
},
{
"epoch": 1.2411569464201042,
"grad_norm": 2.283247232437134,
"learning_rate": 1.4565104614320065e-05,
"loss": 0.7752,
"step": 409
},
{
"epoch": 1.2441915599810336,
"grad_norm": 2.299151659011841,
"learning_rate": 1.453359940302344e-05,
"loss": 0.7794,
"step": 410
},
{
"epoch": 1.247226173541963,
"grad_norm": 2.2340760231018066,
"learning_rate": 1.4502037448176734e-05,
"loss": 0.7811,
"step": 411
},
{
"epoch": 1.2502607871028923,
"grad_norm": 2.305233955383301,
"learning_rate": 1.4470419144816483e-05,
"loss": 0.7663,
"step": 412
},
{
"epoch": 1.2532954006638217,
"grad_norm": 2.1460888385772705,
"learning_rate": 1.4438744888684481e-05,
"loss": 0.7584,
"step": 413
},
{
"epoch": 1.256330014224751,
"grad_norm": 2.3122851848602295,
"learning_rate": 1.4407015076222845e-05,
"loss": 0.7817,
"step": 414
},
{
"epoch": 1.2593646277856805,
"grad_norm": 2.0583643913269043,
"learning_rate": 1.4375230104569044e-05,
"loss": 0.7695,
"step": 415
},
{
"epoch": 1.2623992413466096,
"grad_norm": 2.262274980545044,
"learning_rate": 1.4343390371550936e-05,
"loss": 0.7739,
"step": 416
},
{
"epoch": 1.2654338549075392,
"grad_norm": 2.151893377304077,
"learning_rate": 1.4311496275681785e-05,
"loss": 0.7789,
"step": 417
},
{
"epoch": 1.2684684684684684,
"grad_norm": 2.28367280960083,
"learning_rate": 1.4279548216155265e-05,
"loss": 0.775,
"step": 418
},
{
"epoch": 1.2715030820293978,
"grad_norm": 2.3059751987457275,
"learning_rate": 1.424754659284048e-05,
"loss": 0.7613,
"step": 419
},
{
"epoch": 1.2745376955903271,
"grad_norm": 2.436896800994873,
"learning_rate": 1.4215491806276944e-05,
"loss": 0.7835,
"step": 420
},
{
"epoch": 1.2775723091512565,
"grad_norm": 2.102220058441162,
"learning_rate": 1.418338425766958e-05,
"loss": 0.7932,
"step": 421
},
{
"epoch": 1.280606922712186,
"grad_norm": 2.1711723804473877,
"learning_rate": 1.4151224348883692e-05,
"loss": 0.7668,
"step": 422
},
{
"epoch": 1.2836415362731153,
"grad_norm": 2.213289737701416,
"learning_rate": 1.4119012482439929e-05,
"loss": 0.7745,
"step": 423
},
{
"epoch": 1.2866761498340447,
"grad_norm": 2.0960137844085693,
"learning_rate": 1.408674906150926e-05,
"loss": 0.7742,
"step": 424
},
{
"epoch": 1.2897107633949738,
"grad_norm": 2.081200122833252,
"learning_rate": 1.4054434489907916e-05,
"loss": 0.7652,
"step": 425
},
{
"epoch": 1.2927453769559032,
"grad_norm": 2.284423351287842,
"learning_rate": 1.4022069172092354e-05,
"loss": 0.7762,
"step": 426
},
{
"epoch": 1.2957799905168326,
"grad_norm": 2.292639970779419,
"learning_rate": 1.3989653513154165e-05,
"loss": 0.7644,
"step": 427
},
{
"epoch": 1.298814604077762,
"grad_norm": 2.282759428024292,
"learning_rate": 1.3957187918815032e-05,
"loss": 0.7658,
"step": 428
},
{
"epoch": 1.3018492176386913,
"grad_norm": 2.2170190811157227,
"learning_rate": 1.3924672795421638e-05,
"loss": 0.7661,
"step": 429
},
{
"epoch": 1.3048838311996207,
"grad_norm": 2.202991247177124,
"learning_rate": 1.3892108549940583e-05,
"loss": 0.7881,
"step": 430
},
{
"epoch": 1.30791844476055,
"grad_norm": 2.148986339569092,
"learning_rate": 1.3859495589953289e-05,
"loss": 0.7865,
"step": 431
},
{
"epoch": 1.3109530583214792,
"grad_norm": 2.1265058517456055,
"learning_rate": 1.3826834323650899e-05,
"loss": 0.789,
"step": 432
},
{
"epoch": 1.3139876718824088,
"grad_norm": 2.1798737049102783,
"learning_rate": 1.3794125159829173e-05,
"loss": 0.7707,
"step": 433
},
{
"epoch": 1.317022285443338,
"grad_norm": 2.2072978019714355,
"learning_rate": 1.376136850788336e-05,
"loss": 0.763,
"step": 434
},
{
"epoch": 1.3200568990042674,
"grad_norm": 2.12349534034729,
"learning_rate": 1.3728564777803089e-05,
"loss": 0.7505,
"step": 435
},
{
"epoch": 1.3230915125651967,
"grad_norm": 2.104276180267334,
"learning_rate": 1.3695714380167221e-05,
"loss": 0.7891,
"step": 436
},
{
"epoch": 1.3261261261261261,
"grad_norm": 2.038515090942383,
"learning_rate": 1.3662817726138729e-05,
"loss": 0.7668,
"step": 437
},
{
"epoch": 1.3291607396870555,
"grad_norm": 2.0557668209075928,
"learning_rate": 1.3629875227459532e-05,
"loss": 0.7685,
"step": 438
},
{
"epoch": 1.3321953532479849,
"grad_norm": 2.221299886703491,
"learning_rate": 1.359688729644536e-05,
"loss": 0.7765,
"step": 439
},
{
"epoch": 1.3352299668089143,
"grad_norm": 2.383873462677002,
"learning_rate": 1.356385434598057e-05,
"loss": 0.7863,
"step": 440
},
{
"epoch": 1.3382645803698434,
"grad_norm": 2.144969940185547,
"learning_rate": 1.3530776789513009e-05,
"loss": 0.7854,
"step": 441
},
{
"epoch": 1.341299193930773,
"grad_norm": 2.3431177139282227,
"learning_rate": 1.3497655041048812e-05,
"loss": 0.7491,
"step": 442
},
{
"epoch": 1.3443338074917022,
"grad_norm": 2.1558756828308105,
"learning_rate": 1.3464489515147239e-05,
"loss": 0.7935,
"step": 443
},
{
"epoch": 1.3473684210526315,
"grad_norm": 2.4032328128814697,
"learning_rate": 1.3431280626915466e-05,
"loss": 0.765,
"step": 444
},
{
"epoch": 1.350403034613561,
"grad_norm": 2.373549461364746,
"learning_rate": 1.3398028792003413e-05,
"loss": 0.766,
"step": 445
},
{
"epoch": 1.3534376481744903,
"grad_norm": 2.38440203666687,
"learning_rate": 1.3364734426598527e-05,
"loss": 0.7849,
"step": 446
},
{
"epoch": 1.3564722617354197,
"grad_norm": 2.319101333618164,
"learning_rate": 1.3331397947420578e-05,
"loss": 0.7738,
"step": 447
},
{
"epoch": 1.359506875296349,
"grad_norm": 2.1911866664886475,
"learning_rate": 1.3298019771716435e-05,
"loss": 0.7779,
"step": 448
},
{
"epoch": 1.3625414888572784,
"grad_norm": 2.273451328277588,
"learning_rate": 1.3264600317254854e-05,
"loss": 0.76,
"step": 449
},
{
"epoch": 1.3655761024182076,
"grad_norm": 2.2576355934143066,
"learning_rate": 1.3231140002321252e-05,
"loss": 0.7687,
"step": 450
},
{
"epoch": 1.3655761024182076,
"eval_loss": NaN,
"eval_runtime": 203.9814,
"eval_samples_per_second": 20.384,
"eval_steps_per_second": 5.099,
"step": 450
},
{
"epoch": 1.368610715979137,
"grad_norm": 2.2846243381500244,
"learning_rate": 1.3197639245712454e-05,
"loss": 0.77,
"step": 451
},
{
"epoch": 1.3716453295400663,
"grad_norm": 2.1583635807037354,
"learning_rate": 1.3164098466731467e-05,
"loss": 0.7681,
"step": 452
},
{
"epoch": 1.3746799431009957,
"grad_norm": 2.126995325088501,
"learning_rate": 1.3130518085182224e-05,
"loss": 0.7755,
"step": 453
},
{
"epoch": 1.377714556661925,
"grad_norm": 2.2357747554779053,
"learning_rate": 1.3096898521364338e-05,
"loss": 0.7509,
"step": 454
},
{
"epoch": 1.3807491702228545,
"grad_norm": 2.0657386779785156,
"learning_rate": 1.3063240196067837e-05,
"loss": 0.8043,
"step": 455
},
{
"epoch": 1.3837837837837839,
"grad_norm": 2.1653802394866943,
"learning_rate": 1.3029543530567884e-05,
"loss": 0.7676,
"step": 456
},
{
"epoch": 1.386818397344713,
"grad_norm": 2.161508798599243,
"learning_rate": 1.2995808946619533e-05,
"loss": 0.7735,
"step": 457
},
{
"epoch": 1.3898530109056426,
"grad_norm": 2.185350179672241,
"learning_rate": 1.2962036866452423e-05,
"loss": 0.7891,
"step": 458
},
{
"epoch": 1.3928876244665718,
"grad_norm": 2.198807954788208,
"learning_rate": 1.2928227712765504e-05,
"loss": 0.7657,
"step": 459
},
{
"epoch": 1.3959222380275011,
"grad_norm": 2.2574117183685303,
"learning_rate": 1.2894381908721757e-05,
"loss": 0.7893,
"step": 460
},
{
"epoch": 1.3989568515884305,
"grad_norm": 2.2577221393585205,
"learning_rate": 1.2860499877942876e-05,
"loss": 0.7787,
"step": 461
},
{
"epoch": 1.40199146514936,
"grad_norm": 2.265421152114868,
"learning_rate": 1.282658204450398e-05,
"loss": 0.7714,
"step": 462
},
{
"epoch": 1.4050260787102893,
"grad_norm": 2.0731489658355713,
"learning_rate": 1.2792628832928302e-05,
"loss": 0.782,
"step": 463
},
{
"epoch": 1.4080606922712187,
"grad_norm": 2.130533218383789,
"learning_rate": 1.275864066818188e-05,
"loss": 0.7873,
"step": 464
},
{
"epoch": 1.411095305832148,
"grad_norm": 2.0858917236328125,
"learning_rate": 1.2724617975668229e-05,
"loss": 0.8005,
"step": 465
},
{
"epoch": 1.4141299193930772,
"grad_norm": 2.0890703201293945,
"learning_rate": 1.2690561181223024e-05,
"loss": 0.775,
"step": 466
},
{
"epoch": 1.4171645329540066,
"grad_norm": 2.3934412002563477,
"learning_rate": 1.2656470711108763e-05,
"loss": 0.7931,
"step": 467
},
{
"epoch": 1.420199146514936,
"grad_norm": 2.2045512199401855,
"learning_rate": 1.2622346992009447e-05,
"loss": 0.7747,
"step": 468
},
{
"epoch": 1.4232337600758653,
"grad_norm": 2.371346950531006,
"learning_rate": 1.2588190451025209e-05,
"loss": 0.7581,
"step": 469
},
{
"epoch": 1.4262683736367947,
"grad_norm": 2.1523914337158203,
"learning_rate": 1.2554001515667009e-05,
"loss": 0.7741,
"step": 470
},
{
"epoch": 1.429302987197724,
"grad_norm": 2.2948853969573975,
"learning_rate": 1.2519780613851254e-05,
"loss": 0.7925,
"step": 471
},
{
"epoch": 1.4323376007586535,
"grad_norm": 2.2751922607421875,
"learning_rate": 1.2485528173894447e-05,
"loss": 0.7784,
"step": 472
},
{
"epoch": 1.4353722143195826,
"grad_norm": 2.1596484184265137,
"learning_rate": 1.2451244624507831e-05,
"loss": 0.7895,
"step": 473
},
{
"epoch": 1.4384068278805122,
"grad_norm": 2.2292542457580566,
"learning_rate": 1.2416930394792026e-05,
"loss": 0.7698,
"step": 474
},
{
"epoch": 1.4414414414414414,
"grad_norm": 2.1559245586395264,
"learning_rate": 1.238258591423165e-05,
"loss": 0.776,
"step": 475
},
{
"epoch": 1.4444760550023708,
"grad_norm": 2.072768211364746,
"learning_rate": 1.234821161268995e-05,
"loss": 0.7665,
"step": 476
},
{
"epoch": 1.4475106685633001,
"grad_norm": 2.3749032020568848,
"learning_rate": 1.2313807920403419e-05,
"loss": 0.7765,
"step": 477
},
{
"epoch": 1.4505452821242295,
"grad_norm": 2.1614534854888916,
"learning_rate": 1.22793752679764e-05,
"loss": 0.7908,
"step": 478
},
{
"epoch": 1.4535798956851589,
"grad_norm": 2.2183918952941895,
"learning_rate": 1.2244914086375726e-05,
"loss": 0.7662,
"step": 479
},
{
"epoch": 1.4566145092460883,
"grad_norm": 2.230243444442749,
"learning_rate": 1.22104248069253e-05,
"loss": 0.7758,
"step": 480
},
{
"epoch": 1.4596491228070176,
"grad_norm": 2.503591775894165,
"learning_rate": 1.2175907861300698e-05,
"loss": 0.7739,
"step": 481
},
{
"epoch": 1.4626837363679468,
"grad_norm": 2.4481165409088135,
"learning_rate": 1.2141363681523777e-05,
"loss": 0.788,
"step": 482
},
{
"epoch": 1.4657183499288762,
"grad_norm": 2.30068302154541,
"learning_rate": 1.2106792699957264e-05,
"loss": 0.7905,
"step": 483
},
{
"epoch": 1.4687529634898056,
"grad_norm": 2.2382349967956543,
"learning_rate": 1.2072195349299344e-05,
"loss": 0.7617,
"step": 484
},
{
"epoch": 1.471787577050735,
"grad_norm": 2.2054882049560547,
"learning_rate": 1.2037572062578238e-05,
"loss": 0.7802,
"step": 485
},
{
"epoch": 1.4748221906116643,
"grad_norm": 2.282318115234375,
"learning_rate": 1.2002923273146793e-05,
"loss": 0.7711,
"step": 486
},
{
"epoch": 1.4778568041725937,
"grad_norm": 2.1077611446380615,
"learning_rate": 1.1968249414677055e-05,
"loss": 0.761,
"step": 487
},
{
"epoch": 1.480891417733523,
"grad_norm": 2.1558871269226074,
"learning_rate": 1.1933550921154836e-05,
"loss": 0.7872,
"step": 488
},
{
"epoch": 1.4839260312944522,
"grad_norm": 2.332897901535034,
"learning_rate": 1.1898828226874284e-05,
"loss": 0.7791,
"step": 489
},
{
"epoch": 1.4869606448553818,
"grad_norm": 2.0427420139312744,
"learning_rate": 1.1864081766432457e-05,
"loss": 0.782,
"step": 490
},
{
"epoch": 1.489995258416311,
"grad_norm": 2.2689926624298096,
"learning_rate": 1.1829311974723868e-05,
"loss": 0.7659,
"step": 491
},
{
"epoch": 1.4930298719772404,
"grad_norm": 2.2432594299316406,
"learning_rate": 1.1794519286935056e-05,
"loss": 0.7543,
"step": 492
},
{
"epoch": 1.4960644855381697,
"grad_norm": 2.0946452617645264,
"learning_rate": 1.1759704138539121e-05,
"loss": 0.7712,
"step": 493
},
{
"epoch": 1.499099099099099,
"grad_norm": 2.1154541969299316,
"learning_rate": 1.1724866965290302e-05,
"loss": 0.7732,
"step": 494
},
{
"epoch": 1.5021337126600285,
"grad_norm": 2.195223569869995,
"learning_rate": 1.1690008203218493e-05,
"loss": 0.7596,
"step": 495
},
{
"epoch": 1.5051683262209576,
"grad_norm": 2.105151653289795,
"learning_rate": 1.1655128288623803e-05,
"loss": 0.775,
"step": 496
},
{
"epoch": 1.5082029397818872,
"grad_norm": 2.196159601211548,
"learning_rate": 1.1620227658071088e-05,
"loss": 0.7893,
"step": 497
},
{
"epoch": 1.5112375533428164,
"grad_norm": 2.3409769535064697,
"learning_rate": 1.158530674838449e-05,
"loss": 0.7494,
"step": 498
},
{
"epoch": 1.514272166903746,
"grad_norm": 2.177128553390503,
"learning_rate": 1.155036599664198e-05,
"loss": 0.7595,
"step": 499
},
{
"epoch": 1.5173067804646752,
"grad_norm": 2.3704683780670166,
"learning_rate": 1.1515405840169861e-05,
"loss": 0.7607,
"step": 500
},
{
"epoch": 1.5173067804646752,
"eval_loss": NaN,
"eval_runtime": 204.5697,
"eval_samples_per_second": 20.326,
"eval_steps_per_second": 5.084,
"step": 500
},
{
"epoch": 1.5203413940256045,
"grad_norm": 2.175325393676758,
"learning_rate": 1.1480426716537316e-05,
"loss": 0.7626,
"step": 501
},
{
"epoch": 1.523376007586534,
"grad_norm": 2.093395471572876,
"learning_rate": 1.1445429063550925e-05,
"loss": 0.7787,
"step": 502
},
{
"epoch": 1.5264106211474633,
"grad_norm": 2.2766175270080566,
"learning_rate": 1.1410413319249193e-05,
"loss": 0.7592,
"step": 503
},
{
"epoch": 1.5294452347083927,
"grad_norm": 2.00833797454834,
"learning_rate": 1.1375379921897052e-05,
"loss": 0.7685,
"step": 504
},
{
"epoch": 1.5324798482693218,
"grad_norm": 2.2018191814422607,
"learning_rate": 1.1340329309980379e-05,
"loss": 0.753,
"step": 505
},
{
"epoch": 1.5355144618302514,
"grad_norm": 2.228724479675293,
"learning_rate": 1.130526192220052e-05,
"loss": 0.7687,
"step": 506
},
{
"epoch": 1.5385490753911806,
"grad_norm": 2.241725444793701,
"learning_rate": 1.1270178197468788e-05,
"loss": 0.7674,
"step": 507
},
{
"epoch": 1.54158368895211,
"grad_norm": 2.1697564125061035,
"learning_rate": 1.1235078574900984e-05,
"loss": 0.7726,
"step": 508
},
{
"epoch": 1.5446183025130393,
"grad_norm": 2.197449207305908,
"learning_rate": 1.119996349381187e-05,
"loss": 0.7672,
"step": 509
},
{
"epoch": 1.5476529160739687,
"grad_norm": 2.0595834255218506,
"learning_rate": 1.1164833393709707e-05,
"loss": 0.7706,
"step": 510
},
{
"epoch": 1.550687529634898,
"grad_norm": 2.3002591133117676,
"learning_rate": 1.112968871429073e-05,
"loss": 0.7875,
"step": 511
},
{
"epoch": 1.5537221431958272,
"grad_norm": 2.175219774246216,
"learning_rate": 1.1094529895433653e-05,
"loss": 0.7809,
"step": 512
},
{
"epoch": 1.5567567567567568,
"grad_norm": 2.1368846893310547,
"learning_rate": 1.1059357377194161e-05,
"loss": 0.7878,
"step": 513
},
{
"epoch": 1.559791370317686,
"grad_norm": 2.210344076156616,
"learning_rate": 1.102417159979941e-05,
"loss": 0.7543,
"step": 514
},
{
"epoch": 1.5628259838786156,
"grad_norm": 2.1888577938079834,
"learning_rate": 1.09889730036425e-05,
"loss": 0.7731,
"step": 515
},
{
"epoch": 1.5658605974395448,
"grad_norm": 2.3701112270355225,
"learning_rate": 1.0953762029276982e-05,
"loss": 0.7689,
"step": 516
},
{
"epoch": 1.5688952110004741,
"grad_norm": 2.069556713104248,
"learning_rate": 1.0918539117411334e-05,
"loss": 0.767,
"step": 517
},
{
"epoch": 1.5719298245614035,
"grad_norm": 2.209773302078247,
"learning_rate": 1.0883304708903441e-05,
"loss": 0.7696,
"step": 518
},
{
"epoch": 1.5749644381223329,
"grad_norm": 2.1156795024871826,
"learning_rate": 1.0848059244755093e-05,
"loss": 0.7827,
"step": 519
},
{
"epoch": 1.5779990516832623,
"grad_norm": 2.359513521194458,
"learning_rate": 1.0812803166106445e-05,
"loss": 0.7612,
"step": 520
},
{
"epoch": 1.5810336652441914,
"grad_norm": 2.2499759197235107,
"learning_rate": 1.0777536914230509e-05,
"loss": 0.7671,
"step": 521
},
{
"epoch": 1.584068278805121,
"grad_norm": 2.219525098800659,
"learning_rate": 1.0742260930527625e-05,
"loss": 0.776,
"step": 522
},
{
"epoch": 1.5871028923660502,
"grad_norm": 2.339210033416748,
"learning_rate": 1.0706975656519946e-05,
"loss": 0.7669,
"step": 523
},
{
"epoch": 1.5901375059269798,
"grad_norm": 2.3282480239868164,
"learning_rate": 1.06716815338459e-05,
"loss": 0.7843,
"step": 524
},
{
"epoch": 1.593172119487909,
"grad_norm": 2.09635329246521,
"learning_rate": 1.0636379004254665e-05,
"loss": 0.7598,
"step": 525
},
{
"epoch": 1.5962067330488383,
"grad_norm": 2.3128199577331543,
"learning_rate": 1.0601068509600642e-05,
"loss": 0.7673,
"step": 526
},
{
"epoch": 1.5992413466097677,
"grad_norm": 2.3669700622558594,
"learning_rate": 1.0565750491837925e-05,
"loss": 0.7697,
"step": 527
},
{
"epoch": 1.602275960170697,
"grad_norm": 2.2540953159332275,
"learning_rate": 1.0530425393014773e-05,
"loss": 0.7641,
"step": 528
},
{
"epoch": 1.6053105737316264,
"grad_norm": 2.3089439868927,
"learning_rate": 1.049509365526807e-05,
"loss": 0.768,
"step": 529
},
{
"epoch": 1.6083451872925556,
"grad_norm": 2.0084259510040283,
"learning_rate": 1.0459755720817797e-05,
"loss": 0.7504,
"step": 530
},
{
"epoch": 1.6113798008534852,
"grad_norm": 2.261798143386841,
"learning_rate": 1.0424412031961485e-05,
"loss": 0.7629,
"step": 531
},
{
"epoch": 1.6144144144144144,
"grad_norm": 2.2108161449432373,
"learning_rate": 1.0389063031068698e-05,
"loss": 0.738,
"step": 532
},
{
"epoch": 1.6174490279753437,
"grad_norm": 2.2087652683258057,
"learning_rate": 1.0353709160575488e-05,
"loss": 0.7758,
"step": 533
},
{
"epoch": 1.6204836415362731,
"grad_norm": 2.218838691711426,
"learning_rate": 1.0318350862978848e-05,
"loss": 0.7641,
"step": 534
},
{
"epoch": 1.6235182550972025,
"grad_norm": 2.156991720199585,
"learning_rate": 1.0282988580831183e-05,
"loss": 0.7577,
"step": 535
},
{
"epoch": 1.6265528686581319,
"grad_norm": 2.1950886249542236,
"learning_rate": 1.0247622756734775e-05,
"loss": 0.7888,
"step": 536
},
{
"epoch": 1.629587482219061,
"grad_norm": 2.114649534225464,
"learning_rate": 1.0212253833336237e-05,
"loss": 0.7766,
"step": 537
},
{
"epoch": 1.6326220957799906,
"grad_norm": 2.217928409576416,
"learning_rate": 1.0176882253320968e-05,
"loss": 0.7529,
"step": 538
},
{
"epoch": 1.6356567093409198,
"grad_norm": 2.2367630004882812,
"learning_rate": 1.0141508459407622e-05,
"loss": 0.7699,
"step": 539
},
{
"epoch": 1.6386913229018494,
"grad_norm": 2.1163787841796875,
"learning_rate": 1.0106132894342564e-05,
"loss": 0.7637,
"step": 540
},
{
"epoch": 1.6417259364627785,
"grad_norm": 2.083292245864868,
"learning_rate": 1.0070756000894321e-05,
"loss": 0.783,
"step": 541
},
{
"epoch": 1.644760550023708,
"grad_norm": 2.1091578006744385,
"learning_rate": 1.0035378221848053e-05,
"loss": 0.7848,
"step": 542
},
{
"epoch": 1.6477951635846373,
"grad_norm": 2.0994367599487305,
"learning_rate": 1e-05,
"loss": 0.7698,
"step": 543
},
{
"epoch": 1.6508297771455667,
"grad_norm": 2.2539420127868652,
"learning_rate": 9.964621778151947e-06,
"loss": 0.769,
"step": 544
},
{
"epoch": 1.653864390706496,
"grad_norm": 2.154353380203247,
"learning_rate": 9.929243999105682e-06,
"loss": 0.7496,
"step": 545
},
{
"epoch": 1.6568990042674252,
"grad_norm": 2.265385150909424,
"learning_rate": 9.89386710565744e-06,
"loss": 0.768,
"step": 546
},
{
"epoch": 1.6599336178283548,
"grad_norm": 2.073289632797241,
"learning_rate": 9.858491540592383e-06,
"loss": 0.7807,
"step": 547
},
{
"epoch": 1.662968231389284,
"grad_norm": 2.135737657546997,
"learning_rate": 9.823117746679034e-06,
"loss": 0.766,
"step": 548
},
{
"epoch": 1.6660028449502133,
"grad_norm": 2.1423635482788086,
"learning_rate": 9.787746166663765e-06,
"loss": 0.7588,
"step": 549
},
{
"epoch": 1.6690374585111427,
"grad_norm": 2.0772769451141357,
"learning_rate": 9.752377243265229e-06,
"loss": 0.7647,
"step": 550
},
{
"epoch": 1.6690374585111427,
"eval_loss": NaN,
"eval_runtime": 204.4783,
"eval_samples_per_second": 20.335,
"eval_steps_per_second": 5.086,
"step": 550
},
{
"epoch": 1.672072072072072,
"grad_norm": 2.118854284286499,
"learning_rate": 9.71701141916882e-06,
"loss": 0.7728,
"step": 551
},
{
"epoch": 1.6751066856330015,
"grad_norm": 2.1113169193267822,
"learning_rate": 9.681649137021158e-06,
"loss": 0.7796,
"step": 552
},
{
"epoch": 1.6781412991939306,
"grad_norm": 2.064993143081665,
"learning_rate": 9.646290839424515e-06,
"loss": 0.762,
"step": 553
},
{
"epoch": 1.6811759127548602,
"grad_norm": 2.151028633117676,
"learning_rate": 9.610936968931302e-06,
"loss": 0.7709,
"step": 554
},
{
"epoch": 1.6842105263157894,
"grad_norm": 2.201429843902588,
"learning_rate": 9.57558796803852e-06,
"loss": 0.7619,
"step": 555
},
{
"epoch": 1.687245139876719,
"grad_norm": 2.065720796585083,
"learning_rate": 9.540244279182206e-06,
"loss": 0.798,
"step": 556
},
{
"epoch": 1.6902797534376481,
"grad_norm": 2.1895217895507812,
"learning_rate": 9.504906344731933e-06,
"loss": 0.7802,
"step": 557
},
{
"epoch": 1.6933143669985775,
"grad_norm": 2.2065324783325195,
"learning_rate": 9.46957460698523e-06,
"loss": 0.7767,
"step": 558
},
{
"epoch": 1.696348980559507,
"grad_norm": 2.17842698097229,
"learning_rate": 9.434249508162076e-06,
"loss": 0.7604,
"step": 559
},
{
"epoch": 1.6993835941204363,
"grad_norm": 2.164586305618286,
"learning_rate": 9.398931490399363e-06,
"loss": 0.756,
"step": 560
},
{
"epoch": 1.7024182076813656,
"grad_norm": 2.235797643661499,
"learning_rate": 9.363620995745337e-06,
"loss": 0.777,
"step": 561
},
{
"epoch": 1.7054528212422948,
"grad_norm": 2.084228038787842,
"learning_rate": 9.328318466154102e-06,
"loss": 0.765,
"step": 562
},
{
"epoch": 1.7084874348032244,
"grad_norm": 2.1382224559783936,
"learning_rate": 9.293024343480056e-06,
"loss": 0.7712,
"step": 563
},
{
"epoch": 1.7115220483641536,
"grad_norm": 2.304201126098633,
"learning_rate": 9.257739069472375e-06,
"loss": 0.7748,
"step": 564
},
{
"epoch": 1.714556661925083,
"grad_norm": 2.1342921257019043,
"learning_rate": 9.222463085769495e-06,
"loss": 0.762,
"step": 565
},
{
"epoch": 1.7175912754860123,
"grad_norm": 2.189845323562622,
"learning_rate": 9.187196833893559e-06,
"loss": 0.7605,
"step": 566
},
{
"epoch": 1.7206258890469417,
"grad_norm": 2.3078930377960205,
"learning_rate": 9.151940755244912e-06,
"loss": 0.7677,
"step": 567
},
{
"epoch": 1.723660502607871,
"grad_norm": 2.303213357925415,
"learning_rate": 9.11669529109656e-06,
"loss": 0.7735,
"step": 568
},
{
"epoch": 1.7266951161688002,
"grad_norm": 2.3082435131073,
"learning_rate": 9.081460882588668e-06,
"loss": 0.7683,
"step": 569
},
{
"epoch": 1.7297297297297298,
"grad_norm": 2.2995338439941406,
"learning_rate": 9.046237970723022e-06,
"loss": 0.752,
"step": 570
},
{
"epoch": 1.732764343290659,
"grad_norm": 2.0647215843200684,
"learning_rate": 9.011026996357504e-06,
"loss": 0.7731,
"step": 571
},
{
"epoch": 1.7357989568515886,
"grad_norm": 2.160496473312378,
"learning_rate": 8.975828400200592e-06,
"loss": 0.7543,
"step": 572
},
{
"epoch": 1.7388335704125177,
"grad_norm": 2.3624351024627686,
"learning_rate": 8.94064262280584e-06,
"loss": 0.774,
"step": 573
},
{
"epoch": 1.7418681839734471,
"grad_norm": 2.165922164916992,
"learning_rate": 8.90547010456635e-06,
"loss": 0.7815,
"step": 574
},
{
"epoch": 1.7449027975343765,
"grad_norm": 2.138921022415161,
"learning_rate": 8.870311285709274e-06,
"loss": 0.7593,
"step": 575
},
{
"epoch": 1.7479374110953059,
"grad_norm": 2.122746706008911,
"learning_rate": 8.835166606290295e-06,
"loss": 0.783,
"step": 576
},
{
"epoch": 1.7509720246562352,
"grad_norm": 2.119318723678589,
"learning_rate": 8.80003650618813e-06,
"loss": 0.7778,
"step": 577
},
{
"epoch": 1.7540066382171644,
"grad_norm": 2.092092990875244,
"learning_rate": 8.76492142509902e-06,
"loss": 0.765,
"step": 578
},
{
"epoch": 1.757041251778094,
"grad_norm": 2.156517267227173,
"learning_rate": 8.729821802531213e-06,
"loss": 0.7583,
"step": 579
},
{
"epoch": 1.7600758653390232,
"grad_norm": 2.2096939086914062,
"learning_rate": 8.694738077799487e-06,
"loss": 0.7698,
"step": 580
},
{
"epoch": 1.7631104788999525,
"grad_norm": 1.981307029724121,
"learning_rate": 8.659670690019626e-06,
"loss": 0.7771,
"step": 581
},
{
"epoch": 1.766145092460882,
"grad_norm": 2.1369874477386475,
"learning_rate": 8.624620078102952e-06,
"loss": 0.777,
"step": 582
},
{
"epoch": 1.7691797060218113,
"grad_norm": 2.1329455375671387,
"learning_rate": 8.58958668075081e-06,
"loss": 0.7748,
"step": 583
},
{
"epoch": 1.7722143195827407,
"grad_norm": 2.1630642414093018,
"learning_rate": 8.554570936449073e-06,
"loss": 0.7675,
"step": 584
},
{
"epoch": 1.7752489331436698,
"grad_norm": 2.1383938789367676,
"learning_rate": 8.519573283462688e-06,
"loss": 0.7702,
"step": 585
},
{
"epoch": 1.7782835467045994,
"grad_norm": 2.001569986343384,
"learning_rate": 8.484594159830142e-06,
"loss": 0.7553,
"step": 586
},
{
"epoch": 1.7813181602655286,
"grad_norm": 2.1266000270843506,
"learning_rate": 8.449634003358022e-06,
"loss": 0.7757,
"step": 587
},
{
"epoch": 1.7843527738264582,
"grad_norm": 2.172466278076172,
"learning_rate": 8.414693251615513e-06,
"loss": 0.7625,
"step": 588
},
{
"epoch": 1.7873873873873873,
"grad_norm": 2.2140755653381348,
"learning_rate": 8.379772341928916e-06,
"loss": 0.7653,
"step": 589
},
{
"epoch": 1.7904220009483167,
"grad_norm": 2.1517114639282227,
"learning_rate": 8.344871711376202e-06,
"loss": 0.7744,
"step": 590
},
{
"epoch": 1.793456614509246,
"grad_norm": 2.1749329566955566,
"learning_rate": 8.309991796781512e-06,
"loss": 0.7663,
"step": 591
},
{
"epoch": 1.7964912280701755,
"grad_norm": 2.2055773735046387,
"learning_rate": 8.2751330347097e-06,
"loss": 0.7768,
"step": 592
},
{
"epoch": 1.7995258416311048,
"grad_norm": 2.117532253265381,
"learning_rate": 8.24029586146088e-06,
"loss": 0.7774,
"step": 593
},
{
"epoch": 1.802560455192034,
"grad_norm": 2.2061290740966797,
"learning_rate": 8.205480713064947e-06,
"loss": 0.7819,
"step": 594
},
{
"epoch": 1.8055950687529636,
"grad_norm": 2.2168619632720947,
"learning_rate": 8.170688025276134e-06,
"loss": 0.7646,
"step": 595
},
{
"epoch": 1.8086296823138928,
"grad_norm": 2.238391637802124,
"learning_rate": 8.135918233567545e-06,
"loss": 0.7782,
"step": 596
},
{
"epoch": 1.8116642958748221,
"grad_norm": 2.105041265487671,
"learning_rate": 8.101171773125716e-06,
"loss": 0.7525,
"step": 597
},
{
"epoch": 1.8146989094357515,
"grad_norm": 2.0234577655792236,
"learning_rate": 8.066449078845168e-06,
"loss": 0.7658,
"step": 598
},
{
"epoch": 1.817733522996681,
"grad_norm": 2.188751220703125,
"learning_rate": 8.031750585322948e-06,
"loss": 0.7616,
"step": 599
},
{
"epoch": 1.8207681365576103,
"grad_norm": 2.1460800170898438,
"learning_rate": 7.99707672685321e-06,
"loss": 0.7617,
"step": 600
},
{
"epoch": 1.8207681365576103,
"eval_loss": NaN,
"eval_runtime": 204.6188,
"eval_samples_per_second": 20.321,
"eval_steps_per_second": 5.083,
"step": 600
},
{
"epoch": 1.8238027501185394,
"grad_norm": 1.964609146118164,
"learning_rate": 7.962427937421763e-06,
"loss": 0.7876,
"step": 601
},
{
"epoch": 1.826837363679469,
"grad_norm": 2.0918426513671875,
"learning_rate": 7.92780465070066e-06,
"loss": 0.7744,
"step": 602
},
{
"epoch": 1.8298719772403982,
"grad_norm": 2.059992790222168,
"learning_rate": 7.89320730004274e-06,
"loss": 0.7665,
"step": 603
},
{
"epoch": 1.8329065908013278,
"grad_norm": 2.0450618267059326,
"learning_rate": 7.858636318476226e-06,
"loss": 0.7636,
"step": 604
},
{
"epoch": 1.835941204362257,
"grad_norm": 2.1291024684906006,
"learning_rate": 7.824092138699307e-06,
"loss": 0.7652,
"step": 605
},
{
"epoch": 1.8389758179231863,
"grad_norm": 2.0631322860717773,
"learning_rate": 7.789575193074703e-06,
"loss": 0.7819,
"step": 606
},
{
"epoch": 1.8420104314841157,
"grad_norm": 2.1132242679595947,
"learning_rate": 7.755085913624274e-06,
"loss": 0.7507,
"step": 607
},
{
"epoch": 1.845045045045045,
"grad_norm": 2.1836376190185547,
"learning_rate": 7.720624732023604e-06,
"loss": 0.7798,
"step": 608
},
{
"epoch": 1.8480796586059745,
"grad_norm": 2.087453842163086,
"learning_rate": 7.686192079596586e-06,
"loss": 0.7817,
"step": 609
},
{
"epoch": 1.8511142721669036,
"grad_norm": 2.0749568939208984,
"learning_rate": 7.651788387310053e-06,
"loss": 0.7893,
"step": 610
},
{
"epoch": 1.8541488857278332,
"grad_norm": 2.085495710372925,
"learning_rate": 7.617414085768352e-06,
"loss": 0.7765,
"step": 611
},
{
"epoch": 1.8571834992887624,
"grad_norm": 2.1710901260375977,
"learning_rate": 7.5830696052079754e-06,
"loss": 0.7745,
"step": 612
},
{
"epoch": 1.860218112849692,
"grad_norm": 2.12780499458313,
"learning_rate": 7.548755375492173e-06,
"loss": 0.7845,
"step": 613
},
{
"epoch": 1.8632527264106211,
"grad_norm": 2.240459680557251,
"learning_rate": 7.514471826105556e-06,
"loss": 0.7731,
"step": 614
},
{
"epoch": 1.8662873399715505,
"grad_norm": 2.130253553390503,
"learning_rate": 7.480219386148751e-06,
"loss": 0.7904,
"step": 615
},
{
"epoch": 1.8693219535324799,
"grad_norm": 2.216806650161743,
"learning_rate": 7.445998484332994e-06,
"loss": 0.7529,
"step": 616
},
{
"epoch": 1.8723565670934093,
"grad_norm": 2.250821352005005,
"learning_rate": 7.411809548974792e-06,
"loss": 0.7566,
"step": 617
},
{
"epoch": 1.8753911806543386,
"grad_norm": 2.088843822479248,
"learning_rate": 7.377653007990559e-06,
"loss": 0.7568,
"step": 618
},
{
"epoch": 1.8784257942152678,
"grad_norm": 2.108356237411499,
"learning_rate": 7.343529288891239e-06,
"loss": 0.7769,
"step": 619
},
{
"epoch": 1.8814604077761974,
"grad_norm": 2.1201224327087402,
"learning_rate": 7.3094388187769815e-06,
"loss": 0.7651,
"step": 620
},
{
"epoch": 1.8844950213371265,
"grad_norm": 2.066652536392212,
"learning_rate": 7.275382024331773e-06,
"loss": 0.7773,
"step": 621
},
{
"epoch": 1.887529634898056,
"grad_norm": 2.1200613975524902,
"learning_rate": 7.241359331818121e-06,
"loss": 0.7835,
"step": 622
},
{
"epoch": 1.8905642484589853,
"grad_norm": 2.160304546356201,
"learning_rate": 7.2073711670717e-06,
"loss": 0.7489,
"step": 623
},
{
"epoch": 1.8935988620199147,
"grad_norm": 2.2403597831726074,
"learning_rate": 7.173417955496025e-06,
"loss": 0.7579,
"step": 624
},
{
"epoch": 1.896633475580844,
"grad_norm": 2.083029270172119,
"learning_rate": 7.13950012205713e-06,
"loss": 0.7715,
"step": 625
},
{
"epoch": 1.8996680891417732,
"grad_norm": 2.031341075897217,
"learning_rate": 7.105618091278246e-06,
"loss": 0.7494,
"step": 626
},
{
"epoch": 1.9027027027027028,
"grad_norm": 2.155595541000366,
"learning_rate": 7.071772287234497e-06,
"loss": 0.7797,
"step": 627
},
{
"epoch": 1.905737316263632,
"grad_norm": 2.2580630779266357,
"learning_rate": 7.037963133547583e-06,
"loss": 0.7801,
"step": 628
},
{
"epoch": 1.9087719298245616,
"grad_norm": 2.2416763305664062,
"learning_rate": 7.004191053380469e-06,
"loss": 0.7753,
"step": 629
},
{
"epoch": 1.9118065433854907,
"grad_norm": 2.0690417289733887,
"learning_rate": 6.970456469432116e-06,
"loss": 0.7693,
"step": 630
},
{
"epoch": 1.91484115694642,
"grad_norm": 2.0331270694732666,
"learning_rate": 6.936759803932167e-06,
"loss": 0.7578,
"step": 631
},
{
"epoch": 1.9178757705073495,
"grad_norm": 2.1094417572021484,
"learning_rate": 6.903101478635663e-06,
"loss": 0.7589,
"step": 632
},
{
"epoch": 1.9209103840682789,
"grad_norm": 2.042919635772705,
"learning_rate": 6.869481914817779e-06,
"loss": 0.7659,
"step": 633
},
{
"epoch": 1.9239449976292082,
"grad_norm": 2.0610294342041016,
"learning_rate": 6.835901533268536e-06,
"loss": 0.7648,
"step": 634
},
{
"epoch": 1.9269796111901374,
"grad_norm": 2.0473015308380127,
"learning_rate": 6.802360754287548e-06,
"loss": 0.7763,
"step": 635
},
{
"epoch": 1.930014224751067,
"grad_norm": 2.1137821674346924,
"learning_rate": 6.768859997678751e-06,
"loss": 0.7849,
"step": 636
},
{
"epoch": 1.9330488383119961,
"grad_norm": 2.12430477142334,
"learning_rate": 6.735399682745145e-06,
"loss": 0.7644,
"step": 637
},
{
"epoch": 1.9360834518729255,
"grad_norm": 2.0620853900909424,
"learning_rate": 6.701980228283569e-06,
"loss": 0.7733,
"step": 638
},
{
"epoch": 1.939118065433855,
"grad_norm": 2.124861478805542,
"learning_rate": 6.668602052579425e-06,
"loss": 0.7538,
"step": 639
},
{
"epoch": 1.9421526789947843,
"grad_norm": 2.003215789794922,
"learning_rate": 6.635265573401475e-06,
"loss": 0.7574,
"step": 640
},
{
"epoch": 1.9451872925557137,
"grad_norm": 2.1591827869415283,
"learning_rate": 6.601971207996592e-06,
"loss": 0.7591,
"step": 641
},
{
"epoch": 1.9482219061166428,
"grad_norm": 2.1029608249664307,
"learning_rate": 6.5687193730845375e-06,
"loss": 0.7927,
"step": 642
},
{
"epoch": 1.9512565196775724,
"grad_norm": 2.069796085357666,
"learning_rate": 6.535510484852767e-06,
"loss": 0.7491,
"step": 643
},
{
"epoch": 1.9542911332385016,
"grad_norm": 2.0876927375793457,
"learning_rate": 6.50234495895119e-06,
"loss": 0.7615,
"step": 644
},
{
"epoch": 1.9573257467994312,
"grad_norm": 2.0911192893981934,
"learning_rate": 6.469223210486992e-06,
"loss": 0.7579,
"step": 645
},
{
"epoch": 1.9603603603603603,
"grad_norm": 2.2714855670928955,
"learning_rate": 6.4361456540194325e-06,
"loss": 0.7543,
"step": 646
},
{
"epoch": 1.9633949739212897,
"grad_norm": 2.075011968612671,
"learning_rate": 6.403112703554643e-06,
"loss": 0.752,
"step": 647
},
{
"epoch": 1.966429587482219,
"grad_norm": 2.097029447555542,
"learning_rate": 6.370124772540469e-06,
"loss": 0.7338,
"step": 648
},
{
"epoch": 1.9694642010431485,
"grad_norm": 2.147951602935791,
"learning_rate": 6.337182273861273e-06,
"loss": 0.7735,
"step": 649
},
{
"epoch": 1.9724988146040778,
"grad_norm": 2.0643298625946045,
"learning_rate": 6.3042856198327795e-06,
"loss": 0.7775,
"step": 650
},
{
"epoch": 1.9724988146040778,
"eval_loss": NaN,
"eval_runtime": 204.7127,
"eval_samples_per_second": 20.311,
"eval_steps_per_second": 5.08,
"step": 650
},
{
"epoch": 1.975533428165007,
"grad_norm": 2.043440580368042,
"learning_rate": 6.2714352221969155e-06,
"loss": 0.7593,
"step": 651
},
{
"epoch": 1.9785680417259366,
"grad_norm": 2.250242233276367,
"learning_rate": 6.238631492116643e-06,
"loss": 0.7657,
"step": 652
},
{
"epoch": 1.9816026552868657,
"grad_norm": 2.068997383117676,
"learning_rate": 6.205874840170833e-06,
"loss": 0.7774,
"step": 653
},
{
"epoch": 1.9846372688477951,
"grad_norm": 2.1658010482788086,
"learning_rate": 6.173165676349103e-06,
"loss": 0.7792,
"step": 654
},
{
"epoch": 1.9876718824087245,
"grad_norm": 2.1947412490844727,
"learning_rate": 6.140504410046712e-06,
"loss": 0.7631,
"step": 655
},
{
"epoch": 1.9907064959696539,
"grad_norm": 2.082087278366089,
"learning_rate": 6.107891450059419e-06,
"loss": 0.7564,
"step": 656
},
{
"epoch": 1.9937411095305833,
"grad_norm": 2.043905019760132,
"learning_rate": 6.075327204578363e-06,
"loss": 0.7834,
"step": 657
},
{
"epoch": 1.9967757230915124,
"grad_norm": 2.1089298725128174,
"learning_rate": 6.042812081184973e-06,
"loss": 0.7595,
"step": 658
},
{
"epoch": 1.999810336652442,
"grad_norm": 2.0419466495513916,
"learning_rate": 6.010346486845837e-06,
"loss": 0.7481,
"step": 659
},
{
"epoch": 2.002844950213371,
"grad_norm": 2.1154048442840576,
"learning_rate": 5.97793082790765e-06,
"loss": 0.7479,
"step": 660
},
{
"epoch": 2.0058795637743008,
"grad_norm": 1.9878915548324585,
"learning_rate": 5.945565510092086e-06,
"loss": 0.7547,
"step": 661
},
{
"epoch": 2.00891417733523,
"grad_norm": 2.1136927604675293,
"learning_rate": 5.913250938490744e-06,
"loss": 0.7294,
"step": 662
},
{
"epoch": 2.0119487908961595,
"grad_norm": 2.1599817276000977,
"learning_rate": 5.880987517560075e-06,
"loss": 0.744,
"step": 663
},
{
"epoch": 2.0149834044570887,
"grad_norm": 2.200779438018799,
"learning_rate": 5.84877565111631e-06,
"loss": 0.7353,
"step": 664
},
{
"epoch": 2.018018018018018,
"grad_norm": 2.05635404586792,
"learning_rate": 5.81661574233042e-06,
"loss": 0.7534,
"step": 665
},
{
"epoch": 2.0210526315789474,
"grad_norm": 2.193406820297241,
"learning_rate": 5.784508193723058e-06,
"loss": 0.7443,
"step": 666
},
{
"epoch": 2.0240872451398766,
"grad_norm": 2.120541572570801,
"learning_rate": 5.752453407159521e-06,
"loss": 0.7414,
"step": 667
},
{
"epoch": 2.027121858700806,
"grad_norm": 2.075017213821411,
"learning_rate": 5.720451783844741e-06,
"loss": 0.7439,
"step": 668
},
{
"epoch": 2.0301564722617353,
"grad_norm": 2.1643104553222656,
"learning_rate": 5.688503724318217e-06,
"loss": 0.7565,
"step": 669
},
{
"epoch": 2.033191085822665,
"grad_norm": 2.148364782333374,
"learning_rate": 5.656609628449064e-06,
"loss": 0.7353,
"step": 670
},
{
"epoch": 2.036225699383594,
"grad_norm": 2.205618143081665,
"learning_rate": 5.6247698954309616e-06,
"loss": 0.7529,
"step": 671
},
{
"epoch": 2.0392603129445233,
"grad_norm": 2.1727452278137207,
"learning_rate": 5.592984923777156e-06,
"loss": 0.7439,
"step": 672
},
{
"epoch": 2.042294926505453,
"grad_norm": 2.1897435188293457,
"learning_rate": 5.561255111315525e-06,
"loss": 0.7578,
"step": 673
},
{
"epoch": 2.045329540066382,
"grad_norm": 2.155768871307373,
"learning_rate": 5.529580855183518e-06,
"loss": 0.7687,
"step": 674
},
{
"epoch": 2.0483641536273116,
"grad_norm": 2.100489616394043,
"learning_rate": 5.497962551823266e-06,
"loss": 0.7333,
"step": 675
},
{
"epoch": 2.0513987671882408,
"grad_norm": 2.0687026977539062,
"learning_rate": 5.466400596976568e-06,
"loss": 0.745,
"step": 676
},
{
"epoch": 2.0544333807491704,
"grad_norm": 2.144482135772705,
"learning_rate": 5.434895385679937e-06,
"loss": 0.7369,
"step": 677
},
{
"epoch": 2.0574679943100995,
"grad_norm": 2.1690595149993896,
"learning_rate": 5.403447312259702e-06,
"loss": 0.7443,
"step": 678
},
{
"epoch": 2.060502607871029,
"grad_norm": 2.0905041694641113,
"learning_rate": 5.3720567703270135e-06,
"loss": 0.7333,
"step": 679
},
{
"epoch": 2.0635372214319583,
"grad_norm": 2.205411911010742,
"learning_rate": 5.340724152772956e-06,
"loss": 0.7399,
"step": 680
},
{
"epoch": 2.0665718349928874,
"grad_norm": 2.3007540702819824,
"learning_rate": 5.3094498517636324e-06,
"loss": 0.7452,
"step": 681
},
{
"epoch": 2.069606448553817,
"grad_norm": 2.2882180213928223,
"learning_rate": 5.278234258735215e-06,
"loss": 0.7285,
"step": 682
},
{
"epoch": 2.072641062114746,
"grad_norm": 2.2169337272644043,
"learning_rate": 5.247077764389099e-06,
"loss": 0.7437,
"step": 683
},
{
"epoch": 2.075675675675676,
"grad_norm": 2.1462368965148926,
"learning_rate": 5.215980758686978e-06,
"loss": 0.7366,
"step": 684
},
{
"epoch": 2.078710289236605,
"grad_norm": 2.200030565261841,
"learning_rate": 5.18494363084596e-06,
"loss": 0.7546,
"step": 685
},
{
"epoch": 2.0817449027975345,
"grad_norm": 2.1369688510894775,
"learning_rate": 5.153966769333734e-06,
"loss": 0.7388,
"step": 686
},
{
"epoch": 2.0847795163584637,
"grad_norm": 2.191958427429199,
"learning_rate": 5.1230505618636575e-06,
"loss": 0.7297,
"step": 687
},
{
"epoch": 2.0878141299193933,
"grad_norm": 2.1999082565307617,
"learning_rate": 5.092195395389937e-06,
"loss": 0.7428,
"step": 688
},
{
"epoch": 2.0908487434803225,
"grad_norm": 2.373140335083008,
"learning_rate": 5.061401656102791e-06,
"loss": 0.7585,
"step": 689
},
{
"epoch": 2.0938833570412516,
"grad_norm": 2.2395219802856445,
"learning_rate": 5.030669729423572e-06,
"loss": 0.7382,
"step": 690
},
{
"epoch": 2.096917970602181,
"grad_norm": 2.1638119220733643,
"learning_rate": 5.000000000000003e-06,
"loss": 0.7538,
"step": 691
},
{
"epoch": 2.0999525841631104,
"grad_norm": 2.173945188522339,
"learning_rate": 4.969392851701306e-06,
"loss": 0.7403,
"step": 692
},
{
"epoch": 2.10298719772404,
"grad_norm": 2.2739264965057373,
"learning_rate": 4.938848667613436e-06,
"loss": 0.7459,
"step": 693
},
{
"epoch": 2.106021811284969,
"grad_norm": 2.2303757667541504,
"learning_rate": 4.908367830034284e-06,
"loss": 0.717,
"step": 694
},
{
"epoch": 2.1090564248458987,
"grad_norm": 2.083385944366455,
"learning_rate": 4.8779507204688595e-06,
"loss": 0.747,
"step": 695
},
{
"epoch": 2.112091038406828,
"grad_norm": 2.252095937728882,
"learning_rate": 4.84759771962455e-06,
"loss": 0.7414,
"step": 696
},
{
"epoch": 2.115125651967757,
"grad_norm": 2.192476272583008,
"learning_rate": 4.817309207406347e-06,
"loss": 0.7078,
"step": 697
},
{
"epoch": 2.1181602655286866,
"grad_norm": 2.170762062072754,
"learning_rate": 4.787085562912076e-06,
"loss": 0.7482,
"step": 698
},
{
"epoch": 2.121194879089616,
"grad_norm": 2.1275007724761963,
"learning_rate": 4.756927164427685e-06,
"loss": 0.7415,
"step": 699
},
{
"epoch": 2.1242294926505454,
"grad_norm": 2.1509687900543213,
"learning_rate": 4.726834389422461e-06,
"loss": 0.7326,
"step": 700
},
{
"epoch": 2.1242294926505454,
"eval_loss": NaN,
"eval_runtime": 204.0742,
"eval_samples_per_second": 20.375,
"eval_steps_per_second": 5.096,
"step": 700
},
{
"epoch": 2.1272641062114745,
"grad_norm": 2.0226032733917236,
"learning_rate": 4.696807614544352e-06,
"loss": 0.7411,
"step": 701
},
{
"epoch": 2.130298719772404,
"grad_norm": 2.1866512298583984,
"learning_rate": 4.666847215615225e-06,
"loss": 0.7394,
"step": 702
},
{
"epoch": 2.1333333333333333,
"grad_norm": 2.2013630867004395,
"learning_rate": 4.636953567626176e-06,
"loss": 0.7356,
"step": 703
},
{
"epoch": 2.1363679468942625,
"grad_norm": 2.4076430797576904,
"learning_rate": 4.607127044732827e-06,
"loss": 0.7501,
"step": 704
},
{
"epoch": 2.139402560455192,
"grad_norm": 2.1296753883361816,
"learning_rate": 4.57736802025065e-06,
"loss": 0.7399,
"step": 705
},
{
"epoch": 2.142437174016121,
"grad_norm": 2.22512149810791,
"learning_rate": 4.5476768666502895e-06,
"loss": 0.7553,
"step": 706
},
{
"epoch": 2.145471787577051,
"grad_norm": 2.178199052810669,
"learning_rate": 4.518053955552903e-06,
"loss": 0.7458,
"step": 707
},
{
"epoch": 2.14850640113798,
"grad_norm": 2.0975606441497803,
"learning_rate": 4.488499657725511e-06,
"loss": 0.7234,
"step": 708
},
{
"epoch": 2.1515410146989096,
"grad_norm": 2.1103413105010986,
"learning_rate": 4.459014343076356e-06,
"loss": 0.7431,
"step": 709
},
{
"epoch": 2.1545756282598387,
"grad_norm": 2.2594456672668457,
"learning_rate": 4.429598380650266e-06,
"loss": 0.7578,
"step": 710
},
{
"epoch": 2.1576102418207683,
"grad_norm": 2.2270963191986084,
"learning_rate": 4.400252138624047e-06,
"loss": 0.7461,
"step": 711
},
{
"epoch": 2.1606448553816975,
"grad_norm": 2.191631555557251,
"learning_rate": 4.370975984301866e-06,
"loss": 0.7425,
"step": 712
},
{
"epoch": 2.1636794689426266,
"grad_norm": 2.124616861343384,
"learning_rate": 4.341770284110655e-06,
"loss": 0.7495,
"step": 713
},
{
"epoch": 2.1667140825035562,
"grad_norm": 2.18497371673584,
"learning_rate": 4.312635403595532e-06,
"loss": 0.7239,
"step": 714
},
{
"epoch": 2.1697486960644854,
"grad_norm": 2.149658441543579,
"learning_rate": 4.283571707415214e-06,
"loss": 0.7325,
"step": 715
},
{
"epoch": 2.172783309625415,
"grad_norm": 2.133171558380127,
"learning_rate": 4.25457955933746e-06,
"loss": 0.7324,
"step": 716
},
{
"epoch": 2.175817923186344,
"grad_norm": 2.15606689453125,
"learning_rate": 4.2256593222345185e-06,
"loss": 0.7273,
"step": 717
},
{
"epoch": 2.1788525367472737,
"grad_norm": 2.237285614013672,
"learning_rate": 4.196811358078585e-06,
"loss": 0.7463,
"step": 718
},
{
"epoch": 2.181887150308203,
"grad_norm": 2.267974853515625,
"learning_rate": 4.168036027937267e-06,
"loss": 0.7405,
"step": 719
},
{
"epoch": 2.1849217638691325,
"grad_norm": 2.0772793292999268,
"learning_rate": 4.139333691969071e-06,
"loss": 0.7418,
"step": 720
},
{
"epoch": 2.1879563774300617,
"grad_norm": 2.1171600818634033,
"learning_rate": 4.1107047094188946e-06,
"loss": 0.7222,
"step": 721
},
{
"epoch": 2.190990990990991,
"grad_norm": 2.3039426803588867,
"learning_rate": 4.082149438613514e-06,
"loss": 0.731,
"step": 722
},
{
"epoch": 2.1940256045519204,
"grad_norm": 2.207404375076294,
"learning_rate": 4.053668236957135e-06,
"loss": 0.7499,
"step": 723
},
{
"epoch": 2.1970602181128496,
"grad_norm": 2.1732053756713867,
"learning_rate": 4.025261460926877e-06,
"loss": 0.7617,
"step": 724
},
{
"epoch": 2.200094831673779,
"grad_norm": 2.537623167037964,
"learning_rate": 3.996929466068344e-06,
"loss": 0.72,
"step": 725
},
{
"epoch": 2.2031294452347083,
"grad_norm": 2.2418384552001953,
"learning_rate": 3.96867260699116e-06,
"loss": 0.727,
"step": 726
},
{
"epoch": 2.206164058795638,
"grad_norm": 2.189655065536499,
"learning_rate": 3.940491237364519e-06,
"loss": 0.7321,
"step": 727
},
{
"epoch": 2.209198672356567,
"grad_norm": 2.1620960235595703,
"learning_rate": 3.912385709912794e-06,
"loss": 0.7421,
"step": 728
},
{
"epoch": 2.2122332859174962,
"grad_norm": 2.216374397277832,
"learning_rate": 3.884356376411089e-06,
"loss": 0.7632,
"step": 729
},
{
"epoch": 2.215267899478426,
"grad_norm": 2.152038335800171,
"learning_rate": 3.8564035876808405e-06,
"loss": 0.7349,
"step": 730
},
{
"epoch": 2.218302513039355,
"grad_norm": 2.1349170207977295,
"learning_rate": 3.828527693585451e-06,
"loss": 0.7373,
"step": 731
},
{
"epoch": 2.2213371266002846,
"grad_norm": 2.1102511882781982,
"learning_rate": 3.8007290430258712e-06,
"loss": 0.7278,
"step": 732
},
{
"epoch": 2.2243717401612138,
"grad_norm": 2.124293088912964,
"learning_rate": 3.7730079839362755e-06,
"loss": 0.7315,
"step": 733
},
{
"epoch": 2.2274063537221434,
"grad_norm": 2.0747625827789307,
"learning_rate": 3.7453648632796745e-06,
"loss": 0.7391,
"step": 734
},
{
"epoch": 2.2304409672830725,
"grad_norm": 2.1584184169769287,
"learning_rate": 3.7178000270435765e-06,
"loss": 0.7526,
"step": 735
},
{
"epoch": 2.2334755808440017,
"grad_norm": 2.0907371044158936,
"learning_rate": 3.690313820235686e-06,
"loss": 0.7176,
"step": 736
},
{
"epoch": 2.2365101944049313,
"grad_norm": 2.100431442260742,
"learning_rate": 3.662906586879542e-06,
"loss": 0.7619,
"step": 737
},
{
"epoch": 2.2395448079658604,
"grad_norm": 2.1690993309020996,
"learning_rate": 3.6355786700102426e-06,
"loss": 0.7385,
"step": 738
},
{
"epoch": 2.24257942152679,
"grad_norm": 2.2165980339050293,
"learning_rate": 3.6083304116701535e-06,
"loss": 0.7577,
"step": 739
},
{
"epoch": 2.245614035087719,
"grad_norm": 2.1825928688049316,
"learning_rate": 3.5811621529045927e-06,
"loss": 0.7353,
"step": 740
},
{
"epoch": 2.2486486486486488,
"grad_norm": 2.1485071182250977,
"learning_rate": 3.5540742337576083e-06,
"loss": 0.7534,
"step": 741
},
{
"epoch": 2.251683262209578,
"grad_norm": 2.193574905395508,
"learning_rate": 3.5270669932676926e-06,
"loss": 0.7395,
"step": 742
},
{
"epoch": 2.2547178757705075,
"grad_norm": 2.152099847793579,
"learning_rate": 3.5001407694635326e-06,
"loss": 0.7265,
"step": 743
},
{
"epoch": 2.2577524893314367,
"grad_norm": 2.2395544052124023,
"learning_rate": 3.4732958993598153e-06,
"loss": 0.7482,
"step": 744
},
{
"epoch": 2.2607871028923663,
"grad_norm": 2.209214448928833,
"learning_rate": 3.446532718952966e-06,
"loss": 0.7366,
"step": 745
},
{
"epoch": 2.2638217164532954,
"grad_norm": 2.282007932662964,
"learning_rate": 3.4198515632169703e-06,
"loss": 0.7641,
"step": 746
},
{
"epoch": 2.2668563300142246,
"grad_norm": 2.1614651679992676,
"learning_rate": 3.3932527660991877e-06,
"loss": 0.7502,
"step": 747
},
{
"epoch": 2.269890943575154,
"grad_norm": 2.306121826171875,
"learning_rate": 3.3667366605161322e-06,
"loss": 0.7464,
"step": 748
},
{
"epoch": 2.2729255571360834,
"grad_norm": 2.2220146656036377,
"learning_rate": 3.340303578349361e-06,
"loss": 0.7497,
"step": 749
},
{
"epoch": 2.275960170697013,
"grad_norm": 2.1184372901916504,
"learning_rate": 3.313953850441266e-06,
"loss": 0.7434,
"step": 750
},
{
"epoch": 2.275960170697013,
"eval_loss": NaN,
"eval_runtime": 203.9554,
"eval_samples_per_second": 20.387,
"eval_steps_per_second": 5.099,
"step": 750
},
{
"epoch": 2.278994784257942,
"grad_norm": 2.2895419597625732,
"learning_rate": 3.2876878065909714e-06,
"loss": 0.7362,
"step": 751
},
{
"epoch": 2.2820293978188717,
"grad_norm": 2.148529291152954,
"learning_rate": 3.2615057755502e-06,
"loss": 0.7558,
"step": 752
},
{
"epoch": 2.285064011379801,
"grad_norm": 2.1223018169403076,
"learning_rate": 3.2354080850191328e-06,
"loss": 0.7591,
"step": 753
},
{
"epoch": 2.28809862494073,
"grad_norm": 2.2450056076049805,
"learning_rate": 3.2093950616423397e-06,
"loss": 0.7335,
"step": 754
},
{
"epoch": 2.2911332385016596,
"grad_norm": 2.173351526260376,
"learning_rate": 3.1834670310046735e-06,
"loss": 0.7546,
"step": 755
},
{
"epoch": 2.2941678520625888,
"grad_norm": 2.0341904163360596,
"learning_rate": 3.157624317627195e-06,
"loss": 0.7447,
"step": 756
},
{
"epoch": 2.2972024656235184,
"grad_norm": 2.0900380611419678,
"learning_rate": 3.1318672449631283e-06,
"loss": 0.7544,
"step": 757
},
{
"epoch": 2.3002370791844475,
"grad_norm": 2.2536418437957764,
"learning_rate": 3.106196135393782e-06,
"loss": 0.7326,
"step": 758
},
{
"epoch": 2.303271692745377,
"grad_norm": 2.1906511783599854,
"learning_rate": 3.0806113102245395e-06,
"loss": 0.7522,
"step": 759
},
{
"epoch": 2.3063063063063063,
"grad_norm": 2.1476638317108154,
"learning_rate": 3.055113089680829e-06,
"loss": 0.7423,
"step": 760
},
{
"epoch": 2.3093409198672354,
"grad_norm": 2.1808698177337646,
"learning_rate": 3.029701792904117e-06,
"loss": 0.7612,
"step": 761
},
{
"epoch": 2.312375533428165,
"grad_norm": 2.085947036743164,
"learning_rate": 3.00437773794791e-06,
"loss": 0.7439,
"step": 762
},
{
"epoch": 2.315410146989094,
"grad_norm": 2.1122984886169434,
"learning_rate": 2.979141241773775e-06,
"loss": 0.7383,
"step": 763
},
{
"epoch": 2.318444760550024,
"grad_norm": 2.1743266582489014,
"learning_rate": 2.953992620247379e-06,
"loss": 0.7458,
"step": 764
},
{
"epoch": 2.321479374110953,
"grad_norm": 2.1469690799713135,
"learning_rate": 2.9289321881345257e-06,
"loss": 0.756,
"step": 765
},
{
"epoch": 2.3245139876718826,
"grad_norm": 2.077439069747925,
"learning_rate": 2.9039602590972203e-06,
"loss": 0.7316,
"step": 766
},
{
"epoch": 2.3275486012328117,
"grad_norm": 2.4407780170440674,
"learning_rate": 2.879077145689746e-06,
"loss": 0.7343,
"step": 767
},
{
"epoch": 2.330583214793741,
"grad_norm": 2.1479241847991943,
"learning_rate": 2.8542831593547483e-06,
"loss": 0.7407,
"step": 768
},
{
"epoch": 2.3336178283546705,
"grad_norm": 2.2073941230773926,
"learning_rate": 2.829578610419337e-06,
"loss": 0.7471,
"step": 769
},
{
"epoch": 2.3366524419155996,
"grad_norm": 2.0292317867279053,
"learning_rate": 2.8049638080912045e-06,
"loss": 0.7332,
"step": 770
},
{
"epoch": 2.339687055476529,
"grad_norm": 2.2681589126586914,
"learning_rate": 2.780439060454756e-06,
"loss": 0.7436,
"step": 771
},
{
"epoch": 2.3427216690374584,
"grad_norm": 2.057676076889038,
"learning_rate": 2.75600467446725e-06,
"loss": 0.7352,
"step": 772
},
{
"epoch": 2.345756282598388,
"grad_norm": 2.2097206115722656,
"learning_rate": 2.7316609559549568e-06,
"loss": 0.7275,
"step": 773
},
{
"epoch": 2.348790896159317,
"grad_norm": 2.3039231300354004,
"learning_rate": 2.707408209609339e-06,
"loss": 0.7556,
"step": 774
},
{
"epoch": 2.3518255097202467,
"grad_norm": 2.097167730331421,
"learning_rate": 2.683246738983217e-06,
"loss": 0.7404,
"step": 775
},
{
"epoch": 2.354860123281176,
"grad_norm": 2.083677053451538,
"learning_rate": 2.6591768464870016e-06,
"loss": 0.7359,
"step": 776
},
{
"epoch": 2.3578947368421055,
"grad_norm": 2.0094263553619385,
"learning_rate": 2.6351988333848787e-06,
"loss": 0.7297,
"step": 777
},
{
"epoch": 2.3609293504030346,
"grad_norm": 2.13173770904541,
"learning_rate": 2.611312999791055e-06,
"loss": 0.7315,
"step": 778
},
{
"epoch": 2.363963963963964,
"grad_norm": 2.1343095302581787,
"learning_rate": 2.587519644666001e-06,
"loss": 0.7309,
"step": 779
},
{
"epoch": 2.3669985775248934,
"grad_norm": 2.122774600982666,
"learning_rate": 2.5638190658126937e-06,
"loss": 0.7428,
"step": 780
},
{
"epoch": 2.3700331910858226,
"grad_norm": 2.1483750343322754,
"learning_rate": 2.5402115598729182e-06,
"loss": 0.7386,
"step": 781
},
{
"epoch": 2.373067804646752,
"grad_norm": 2.189150810241699,
"learning_rate": 2.5166974223235295e-06,
"loss": 0.7237,
"step": 782
},
{
"epoch": 2.3761024182076813,
"grad_norm": 2.156003475189209,
"learning_rate": 2.493276947472756e-06,
"loss": 0.7614,
"step": 783
},
{
"epoch": 2.379137031768611,
"grad_norm": 2.100486993789673,
"learning_rate": 2.4699504284565413e-06,
"loss": 0.7543,
"step": 784
},
{
"epoch": 2.38217164532954,
"grad_norm": 2.1420738697052,
"learning_rate": 2.446718157234832e-06,
"loss": 0.7475,
"step": 785
},
{
"epoch": 2.385206258890469,
"grad_norm": 2.205798625946045,
"learning_rate": 2.4235804245879723e-06,
"loss": 0.7362,
"step": 786
},
{
"epoch": 2.388240872451399,
"grad_norm": 2.193894147872925,
"learning_rate": 2.4005375201130275e-06,
"loss": 0.7456,
"step": 787
},
{
"epoch": 2.391275486012328,
"grad_norm": 2.1630735397338867,
"learning_rate": 2.3775897322201667e-06,
"loss": 0.7553,
"step": 788
},
{
"epoch": 2.3943100995732576,
"grad_norm": 2.247066020965576,
"learning_rate": 2.354737348129077e-06,
"loss": 0.7324,
"step": 789
},
{
"epoch": 2.3973447131341867,
"grad_norm": 2.1332616806030273,
"learning_rate": 2.33198065386533e-06,
"loss": 0.7413,
"step": 790
},
{
"epoch": 2.4003793266951163,
"grad_norm": 2.1438822746276855,
"learning_rate": 2.3093199342568316e-06,
"loss": 0.7215,
"step": 791
},
{
"epoch": 2.4034139402560455,
"grad_norm": 2.1765286922454834,
"learning_rate": 2.2867554729302545e-06,
"loss": 0.7362,
"step": 792
},
{
"epoch": 2.4064485538169746,
"grad_norm": 2.059136390686035,
"learning_rate": 2.2642875523074613e-06,
"loss": 0.7373,
"step": 793
},
{
"epoch": 2.4094831673779042,
"grad_norm": 2.195065975189209,
"learning_rate": 2.2419164536020112e-06,
"loss": 0.7363,
"step": 794
},
{
"epoch": 2.4125177809388334,
"grad_norm": 2.2990760803222656,
"learning_rate": 2.2196424568156073e-06,
"loss": 0.7348,
"step": 795
},
{
"epoch": 2.415552394499763,
"grad_norm": 2.251936435699463,
"learning_rate": 2.197465840734596e-06,
"loss": 0.7257,
"step": 796
},
{
"epoch": 2.418587008060692,
"grad_norm": 2.142132520675659,
"learning_rate": 2.1753868829265046e-06,
"loss": 0.7526,
"step": 797
},
{
"epoch": 2.4216216216216218,
"grad_norm": 2.235971450805664,
"learning_rate": 2.1534058597365284e-06,
"loss": 0.7402,
"step": 798
},
{
"epoch": 2.424656235182551,
"grad_norm": 2.1040003299713135,
"learning_rate": 2.1315230462840985e-06,
"loss": 0.7271,
"step": 799
},
{
"epoch": 2.4276908487434805,
"grad_norm": 2.153122663497925,
"learning_rate": 2.1097387164594406e-06,
"loss": 0.7274,
"step": 800
},
{
"epoch": 2.4276908487434805,
"eval_loss": NaN,
"eval_runtime": 204.5921,
"eval_samples_per_second": 20.323,
"eval_steps_per_second": 5.083,
"step": 800
},
{
"epoch": 2.4307254623044097,
"grad_norm": 2.2538816928863525,
"learning_rate": 2.0880531429201146e-06,
"loss": 0.749,
"step": 801
},
{
"epoch": 2.4337600758653393,
"grad_norm": 2.194967031478882,
"learning_rate": 2.0664665970876496e-06,
"loss": 0.7395,
"step": 802
},
{
"epoch": 2.4367946894262684,
"grad_norm": 2.1779520511627197,
"learning_rate": 2.0449793491441026e-06,
"loss": 0.7476,
"step": 803
},
{
"epoch": 2.4398293029871976,
"grad_norm": 2.243781805038452,
"learning_rate": 2.0235916680287015e-06,
"loss": 0.7268,
"step": 804
},
{
"epoch": 2.442863916548127,
"grad_norm": 2.1497793197631836,
"learning_rate": 2.0023038214344827e-06,
"loss": 0.76,
"step": 805
},
{
"epoch": 2.4458985301090563,
"grad_norm": 2.209360122680664,
"learning_rate": 1.9811160758049163e-06,
"loss": 0.7133,
"step": 806
},
{
"epoch": 2.448933143669986,
"grad_norm": 2.1285481452941895,
"learning_rate": 1.960028696330596e-06,
"loss": 0.7349,
"step": 807
},
{
"epoch": 2.451967757230915,
"grad_norm": 2.16249680519104,
"learning_rate": 1.9390419469459066e-06,
"loss": 0.7392,
"step": 808
},
{
"epoch": 2.4550023707918447,
"grad_norm": 2.1789779663085938,
"learning_rate": 1.9181560903257234e-06,
"loss": 0.7299,
"step": 809
},
{
"epoch": 2.458036984352774,
"grad_norm": 2.196904420852661,
"learning_rate": 1.8973713878821343e-06,
"loss": 0.74,
"step": 810
},
{
"epoch": 2.461071597913703,
"grad_norm": 2.087092399597168,
"learning_rate": 1.8766880997611424e-06,
"loss": 0.7503,
"step": 811
},
{
"epoch": 2.4641062114746326,
"grad_norm": 2.023298978805542,
"learning_rate": 1.8561064848394384e-06,
"loss": 0.7409,
"step": 812
},
{
"epoch": 2.4671408250355618,
"grad_norm": 2.2470693588256836,
"learning_rate": 1.8356268007211442e-06,
"loss": 0.7553,
"step": 813
},
{
"epoch": 2.4701754385964914,
"grad_norm": 2.0922911167144775,
"learning_rate": 1.8152493037345942e-06,
"loss": 0.7257,
"step": 814
},
{
"epoch": 2.4732100521574205,
"grad_norm": 2.100867986679077,
"learning_rate": 1.7949742489291256e-06,
"loss": 0.7275,
"step": 815
},
{
"epoch": 2.47624466571835,
"grad_norm": 2.309476375579834,
"learning_rate": 1.7748018900718856e-06,
"loss": 0.7349,
"step": 816
},
{
"epoch": 2.4792792792792793,
"grad_norm": 2.1686556339263916,
"learning_rate": 1.7547324796446553e-06,
"loss": 0.746,
"step": 817
},
{
"epoch": 2.4823138928402084,
"grad_norm": 2.161126136779785,
"learning_rate": 1.7347662688406907e-06,
"loss": 0.7637,
"step": 818
},
{
"epoch": 2.485348506401138,
"grad_norm": 2.0931811332702637,
"learning_rate": 1.7149035075615795e-06,
"loss": 0.7417,
"step": 819
},
{
"epoch": 2.488383119962067,
"grad_norm": 2.220621347427368,
"learning_rate": 1.6951444444141084e-06,
"loss": 0.7426,
"step": 820
},
{
"epoch": 2.491417733522997,
"grad_norm": 2.1453609466552734,
"learning_rate": 1.6754893267071593e-06,
"loss": 0.7277,
"step": 821
},
{
"epoch": 2.494452347083926,
"grad_norm": 2.204213857650757,
"learning_rate": 1.6559384004486057e-06,
"loss": 0.7456,
"step": 822
},
{
"epoch": 2.4974869606448555,
"grad_norm": 2.1252691745758057,
"learning_rate": 1.6364919103422394e-06,
"loss": 0.7423,
"step": 823
},
{
"epoch": 2.5005215742057847,
"grad_norm": 2.285757064819336,
"learning_rate": 1.6171500997847056e-06,
"loss": 0.7377,
"step": 824
},
{
"epoch": 2.503556187766714,
"grad_norm": 2.1519298553466797,
"learning_rate": 1.5979132108624572e-06,
"loss": 0.7397,
"step": 825
},
{
"epoch": 2.5065908013276434,
"grad_norm": 2.1684153079986572,
"learning_rate": 1.5787814843487226e-06,
"loss": 0.7355,
"step": 826
},
{
"epoch": 2.509625414888573,
"grad_norm": 2.1787893772125244,
"learning_rate": 1.5597551597004968e-06,
"loss": 0.7356,
"step": 827
},
{
"epoch": 2.512660028449502,
"grad_norm": 2.22249436378479,
"learning_rate": 1.5408344750555382e-06,
"loss": 0.7439,
"step": 828
},
{
"epoch": 2.5156946420104314,
"grad_norm": 2.2051734924316406,
"learning_rate": 1.522019667229393e-06,
"loss": 0.7271,
"step": 829
},
{
"epoch": 2.518729255571361,
"grad_norm": 2.0986456871032715,
"learning_rate": 1.5033109717124284e-06,
"loss": 0.7385,
"step": 830
},
{
"epoch": 2.52176386913229,
"grad_norm": 2.115734100341797,
"learning_rate": 1.4847086226668871e-06,
"loss": 0.7317,
"step": 831
},
{
"epoch": 2.5247984826932193,
"grad_norm": 2.02058482170105,
"learning_rate": 1.4662128529239572e-06,
"loss": 0.7524,
"step": 832
},
{
"epoch": 2.527833096254149,
"grad_norm": 2.1982414722442627,
"learning_rate": 1.4478238939808454e-06,
"loss": 0.7409,
"step": 833
},
{
"epoch": 2.5308677098150785,
"grad_norm": 2.1642115116119385,
"learning_rate": 1.4295419759979079e-06,
"loss": 0.7365,
"step": 834
},
{
"epoch": 2.5339023233760076,
"grad_norm": 2.118788003921509,
"learning_rate": 1.4113673277957395e-06,
"loss": 0.7276,
"step": 835
},
{
"epoch": 2.536936936936937,
"grad_norm": 2.0929243564605713,
"learning_rate": 1.393300176852327e-06,
"loss": 0.7449,
"step": 836
},
{
"epoch": 2.5399715504978664,
"grad_norm": 2.1339359283447266,
"learning_rate": 1.3753407493001968e-06,
"loss": 0.7241,
"step": 837
},
{
"epoch": 2.5430061640587955,
"grad_norm": 2.15981125831604,
"learning_rate": 1.3574892699235798e-06,
"loss": 0.7373,
"step": 838
},
{
"epoch": 2.546040777619725,
"grad_norm": 2.184105396270752,
"learning_rate": 1.339745962155613e-06,
"loss": 0.7283,
"step": 839
},
{
"epoch": 2.5490753911806543,
"grad_norm": 2.0236942768096924,
"learning_rate": 1.3221110480755306e-06,
"loss": 0.7279,
"step": 840
},
{
"epoch": 2.552110004741584,
"grad_norm": 2.0451905727386475,
"learning_rate": 1.3045847484058748e-06,
"loss": 0.7417,
"step": 841
},
{
"epoch": 2.555144618302513,
"grad_norm": 2.113330602645874,
"learning_rate": 1.287167282509767e-06,
"loss": 0.7604,
"step": 842
},
{
"epoch": 2.558179231863442,
"grad_norm": 2.0830495357513428,
"learning_rate": 1.2698588683881185e-06,
"loss": 0.7579,
"step": 843
},
{
"epoch": 2.561213845424372,
"grad_norm": 2.065068483352661,
"learning_rate": 1.252659722676941e-06,
"loss": 0.7268,
"step": 844
},
{
"epoch": 2.564248458985301,
"grad_norm": 2.157485008239746,
"learning_rate": 1.2355700606446119e-06,
"loss": 0.7305,
"step": 845
},
{
"epoch": 2.5672830725462306,
"grad_norm": 2.135056257247925,
"learning_rate": 1.2185900961891794e-06,
"loss": 0.7332,
"step": 846
},
{
"epoch": 2.5703176861071597,
"grad_norm": 2.1909451484680176,
"learning_rate": 1.2017200418357077e-06,
"loss": 0.7402,
"step": 847
},
{
"epoch": 2.5733522996680893,
"grad_norm": 2.2838058471679688,
"learning_rate": 1.184960108733586e-06,
"loss": 0.7336,
"step": 848
},
{
"epoch": 2.5763869132290185,
"grad_norm": 2.0348660945892334,
"learning_rate": 1.1683105066539068e-06,
"loss": 0.7367,
"step": 849
},
{
"epoch": 2.5794215267899476,
"grad_norm": 2.088353157043457,
"learning_rate": 1.151771443986842e-06,
"loss": 0.749,
"step": 850
},
{
"epoch": 2.5794215267899476,
"eval_loss": NaN,
"eval_runtime": 205.2076,
"eval_samples_per_second": 20.262,
"eval_steps_per_second": 5.068,
"step": 850
},
{
"epoch": 2.5824561403508772,
"grad_norm": 2.1730659008026123,
"learning_rate": 1.1353431277390125e-06,
"loss": 0.7359,
"step": 851
},
{
"epoch": 2.5854907539118064,
"grad_norm": 2.10697078704834,
"learning_rate": 1.1190257635309276e-06,
"loss": 0.7223,
"step": 852
},
{
"epoch": 2.588525367472736,
"grad_norm": 2.1502668857574463,
"learning_rate": 1.1028195555943877e-06,
"loss": 0.7265,
"step": 853
},
{
"epoch": 2.591559981033665,
"grad_norm": 2.0735018253326416,
"learning_rate": 1.0867247067699315e-06,
"loss": 0.7345,
"step": 854
},
{
"epoch": 2.5945945945945947,
"grad_norm": 2.1325292587280273,
"learning_rate": 1.0707414185043163e-06,
"loss": 0.746,
"step": 855
},
{
"epoch": 2.597629208155524,
"grad_norm": 2.1418328285217285,
"learning_rate": 1.0548698908479671e-06,
"loss": 0.7401,
"step": 856
},
{
"epoch": 2.600663821716453,
"grad_norm": 2.061847448348999,
"learning_rate": 1.0391103224524957e-06,
"loss": 0.7319,
"step": 857
},
{
"epoch": 2.6036984352773827,
"grad_norm": 2.1931376457214355,
"learning_rate": 1.0234629105682104e-06,
"loss": 0.7429,
"step": 858
},
{
"epoch": 2.6067330488383122,
"grad_norm": 2.1396734714508057,
"learning_rate": 1.0079278510416313e-06,
"loss": 0.7369,
"step": 859
},
{
"epoch": 2.6097676623992414,
"grad_norm": 2.2338297367095947,
"learning_rate": 9.925053383130667e-07,
"loss": 0.745,
"step": 860
},
{
"epoch": 2.6128022759601706,
"grad_norm": 2.120897054672241,
"learning_rate": 9.771955654141496e-07,
"loss": 0.7444,
"step": 861
},
{
"epoch": 2.6158368895211,
"grad_norm": 2.204875946044922,
"learning_rate": 9.619987239654405e-07,
"loss": 0.7293,
"step": 862
},
{
"epoch": 2.6188715030820293,
"grad_norm": 2.228181838989258,
"learning_rate": 9.469150041740338e-07,
"loss": 0.7251,
"step": 863
},
{
"epoch": 2.6219061166429585,
"grad_norm": 2.170053720474243,
"learning_rate": 9.319445948311534e-07,
"loss": 0.7444,
"step": 864
},
{
"epoch": 2.624940730203888,
"grad_norm": 2.163527011871338,
"learning_rate": 9.170876833098119e-07,
"loss": 0.7568,
"step": 865
},
{
"epoch": 2.6279753437648177,
"grad_norm": 2.1692657470703125,
"learning_rate": 9.023444555624572e-07,
"loss": 0.736,
"step": 866
},
{
"epoch": 2.631009957325747,
"grad_norm": 2.0737569332122803,
"learning_rate": 8.87715096118642e-07,
"loss": 0.7368,
"step": 867
},
{
"epoch": 2.634044570886676,
"grad_norm": 2.05617094039917,
"learning_rate": 8.731997880827258e-07,
"loss": 0.7334,
"step": 868
},
{
"epoch": 2.6370791844476056,
"grad_norm": 1.9645661115646362,
"learning_rate": 8.587987131315656e-07,
"loss": 0.7422,
"step": 869
},
{
"epoch": 2.6401137980085347,
"grad_norm": 2.0881567001342773,
"learning_rate": 8.445120515122552e-07,
"loss": 0.7296,
"step": 870
},
{
"epoch": 2.6431484115694643,
"grad_norm": 2.2319769859313965,
"learning_rate": 8.303399820398672e-07,
"loss": 0.7502,
"step": 871
},
{
"epoch": 2.6461830251303935,
"grad_norm": 2.1117403507232666,
"learning_rate": 8.162826820952097e-07,
"loss": 0.7319,
"step": 872
},
{
"epoch": 2.649217638691323,
"grad_norm": 2.056861400604248,
"learning_rate": 8.023403276226127e-07,
"loss": 0.7429,
"step": 873
},
{
"epoch": 2.6522522522522523,
"grad_norm": 2.140435218811035,
"learning_rate": 7.885130931277219e-07,
"loss": 0.7418,
"step": 874
},
{
"epoch": 2.6552868658131814,
"grad_norm": 2.11114239692688,
"learning_rate": 7.74801151675314e-07,
"loss": 0.7521,
"step": 875
},
{
"epoch": 2.658321479374111,
"grad_norm": 2.0694892406463623,
"learning_rate": 7.612046748871327e-07,
"loss": 0.7385,
"step": 876
},
{
"epoch": 2.66135609293504,
"grad_norm": 2.155579090118408,
"learning_rate": 7.477238329397419e-07,
"loss": 0.7427,
"step": 877
},
{
"epoch": 2.6643907064959698,
"grad_norm": 2.223646402359009,
"learning_rate": 7.343587945623908e-07,
"loss": 0.731,
"step": 878
},
{
"epoch": 2.667425320056899,
"grad_norm": 2.047801971435547,
"learning_rate": 7.211097270349065e-07,
"loss": 0.7426,
"step": 879
},
{
"epoch": 2.6704599336178285,
"grad_norm": 2.119241714477539,
"learning_rate": 7.07976796185601e-07,
"loss": 0.7407,
"step": 880
},
{
"epoch": 2.6734945471787577,
"grad_norm": 2.2282469272613525,
"learning_rate": 6.949601663891891e-07,
"loss": 0.7541,
"step": 881
},
{
"epoch": 2.676529160739687,
"grad_norm": 2.175909996032715,
"learning_rate": 6.820600005647382e-07,
"loss": 0.7461,
"step": 882
},
{
"epoch": 2.6795637743006164,
"grad_norm": 2.047321081161499,
"learning_rate": 6.692764601736268e-07,
"loss": 0.7197,
"step": 883
},
{
"epoch": 2.682598387861546,
"grad_norm": 2.0702617168426514,
"learning_rate": 6.566097052175213e-07,
"loss": 0.7656,
"step": 884
},
{
"epoch": 2.685633001422475,
"grad_norm": 2.1683459281921387,
"learning_rate": 6.440598942363796e-07,
"loss": 0.7521,
"step": 885
},
{
"epoch": 2.6886676149834043,
"grad_norm": 2.067746162414551,
"learning_rate": 6.316271843064536e-07,
"loss": 0.7531,
"step": 886
},
{
"epoch": 2.691702228544334,
"grad_norm": 2.1310875415802,
"learning_rate": 6.193117310383412e-07,
"loss": 0.7538,
"step": 887
},
{
"epoch": 2.694736842105263,
"grad_norm": 2.1403868198394775,
"learning_rate": 6.071136885750272e-07,
"loss": 0.7373,
"step": 888
},
{
"epoch": 2.6977714556661923,
"grad_norm": 2.049807071685791,
"learning_rate": 5.950332095899547e-07,
"loss": 0.738,
"step": 889
},
{
"epoch": 2.700806069227122,
"grad_norm": 2.080238103866577,
"learning_rate": 5.830704452851166e-07,
"loss": 0.7212,
"step": 890
},
{
"epoch": 2.7038406827880515,
"grad_norm": 2.1021158695220947,
"learning_rate": 5.71225545389158e-07,
"loss": 0.7411,
"step": 891
},
{
"epoch": 2.7068752963489806,
"grad_norm": 2.0948758125305176,
"learning_rate": 5.594986581555173e-07,
"loss": 0.7369,
"step": 892
},
{
"epoch": 2.7099099099099098,
"grad_norm": 2.1047518253326416,
"learning_rate": 5.478899303605512e-07,
"loss": 0.7289,
"step": 893
},
{
"epoch": 2.7129445234708394,
"grad_norm": 2.1207687854766846,
"learning_rate": 5.363995073017047e-07,
"loss": 0.7385,
"step": 894
},
{
"epoch": 2.7159791370317685,
"grad_norm": 2.051896572113037,
"learning_rate": 5.250275327957033e-07,
"loss": 0.7244,
"step": 895
},
{
"epoch": 2.719013750592698,
"grad_norm": 2.220273017883301,
"learning_rate": 5.137741491767345e-07,
"loss": 0.7522,
"step": 896
},
{
"epoch": 2.7220483641536273,
"grad_norm": 2.1100893020629883,
"learning_rate": 5.026394972946813e-07,
"loss": 0.7276,
"step": 897
},
{
"epoch": 2.725082977714557,
"grad_norm": 2.300666332244873,
"learning_rate": 4.91623716513352e-07,
"loss": 0.7404,
"step": 898
},
{
"epoch": 2.728117591275486,
"grad_norm": 2.2343952655792236,
"learning_rate": 4.807269447087348e-07,
"loss": 0.7474,
"step": 899
},
{
"epoch": 2.731152204836415,
"grad_norm": 2.1582469940185547,
"learning_rate": 4.6994931826728094e-07,
"loss": 0.7275,
"step": 900
},
{
"epoch": 2.731152204836415,
"eval_loss": NaN,
"eval_runtime": 204.8855,
"eval_samples_per_second": 20.294,
"eval_steps_per_second": 5.076,
"step": 900
},
{
"epoch": 2.734186818397345,
"grad_norm": 2.057573080062866,
"learning_rate": 4.592909720841843e-07,
"loss": 0.7351,
"step": 901
},
{
"epoch": 2.737221431958274,
"grad_norm": 2.320993423461914,
"learning_rate": 4.487520395617029e-07,
"loss": 0.7086,
"step": 902
},
{
"epoch": 2.7402560455192035,
"grad_norm": 2.2511098384857178,
"learning_rate": 4.3833265260749157e-07,
"loss": 0.723,
"step": 903
},
{
"epoch": 2.7432906590801327,
"grad_norm": 2.182513952255249,
"learning_rate": 4.280329416329365e-07,
"loss": 0.7196,
"step": 904
},
{
"epoch": 2.7463252726410623,
"grad_norm": 2.1579484939575195,
"learning_rate": 4.178530355515409e-07,
"loss": 0.7334,
"step": 905
},
{
"epoch": 2.7493598862019915,
"grad_norm": 2.091196060180664,
"learning_rate": 4.077930617773007e-07,
"loss": 0.7387,
"step": 906
},
{
"epoch": 2.7523944997629206,
"grad_norm": 2.1611766815185547,
"learning_rate": 3.97853146223105e-07,
"loss": 0.7515,
"step": 907
},
{
"epoch": 2.75542911332385,
"grad_norm": 2.240387201309204,
"learning_rate": 3.880334132991792e-07,
"loss": 0.7377,
"step": 908
},
{
"epoch": 2.7584637268847794,
"grad_norm": 2.1546630859375,
"learning_rate": 3.783339859115065e-07,
"loss": 0.7388,
"step": 909
},
{
"epoch": 2.761498340445709,
"grad_norm": 2.0892395973205566,
"learning_rate": 3.687549854603023e-07,
"loss": 0.7488,
"step": 910
},
{
"epoch": 2.764532954006638,
"grad_norm": 2.182509422302246,
"learning_rate": 3.5929653183849444e-07,
"loss": 0.736,
"step": 911
},
{
"epoch": 2.7675675675675677,
"grad_norm": 2.198559522628784,
"learning_rate": 3.49958743430211e-07,
"loss": 0.736,
"step": 912
},
{
"epoch": 2.770602181128497,
"grad_norm": 2.188133955001831,
"learning_rate": 3.4074173710931804e-07,
"loss": 0.7345,
"step": 913
},
{
"epoch": 2.773636794689426,
"grad_norm": 2.133892774581909,
"learning_rate": 3.3164562823793654e-07,
"loss": 0.7275,
"step": 914
},
{
"epoch": 2.7766714082503556,
"grad_norm": 2.159824848175049,
"learning_rate": 3.226705306650113e-07,
"loss": 0.7198,
"step": 915
},
{
"epoch": 2.7797060218112852,
"grad_norm": 2.1087584495544434,
"learning_rate": 3.1381655672488634e-07,
"loss": 0.7381,
"step": 916
},
{
"epoch": 2.7827406353722144,
"grad_norm": 2.0807528495788574,
"learning_rate": 3.050838172358883e-07,
"loss": 0.7386,
"step": 917
},
{
"epoch": 2.7857752489331435,
"grad_norm": 2.1473801136016846,
"learning_rate": 2.9647242149895005e-07,
"loss": 0.7336,
"step": 918
},
{
"epoch": 2.788809862494073,
"grad_norm": 2.219571352005005,
"learning_rate": 2.879824772962381e-07,
"loss": 0.7438,
"step": 919
},
{
"epoch": 2.7918444760550023,
"grad_norm": 2.144059419631958,
"learning_rate": 2.796140908898026e-07,
"loss": 0.7338,
"step": 920
},
{
"epoch": 2.7948790896159315,
"grad_norm": 2.1552734375,
"learning_rate": 2.7136736702025436e-07,
"loss": 0.7345,
"step": 921
},
{
"epoch": 2.797913703176861,
"grad_norm": 2.18730092048645,
"learning_rate": 2.632424089054419e-07,
"loss": 0.7222,
"step": 922
},
{
"epoch": 2.8009483167377907,
"grad_norm": 2.233304500579834,
"learning_rate": 2.552393182391677e-07,
"loss": 0.7474,
"step": 923
},
{
"epoch": 2.80398293029872,
"grad_norm": 2.041757583618164,
"learning_rate": 2.473581951899184e-07,
"loss": 0.7329,
"step": 924
},
{
"epoch": 2.807017543859649,
"grad_norm": 2.1639747619628906,
"learning_rate": 2.395991383995999e-07,
"loss": 0.748,
"step": 925
},
{
"epoch": 2.8100521574205786,
"grad_norm": 2.0612905025482178,
"learning_rate": 2.3196224498231447e-07,
"loss": 0.7311,
"step": 926
},
{
"epoch": 2.8130867709815077,
"grad_norm": 2.260965347290039,
"learning_rate": 2.2444761052313857e-07,
"loss": 0.7353,
"step": 927
},
{
"epoch": 2.8161213845424373,
"grad_norm": 2.1008665561676025,
"learning_rate": 2.1705532907692617e-07,
"loss": 0.7677,
"step": 928
},
{
"epoch": 2.8191559981033665,
"grad_norm": 2.20302414894104,
"learning_rate": 2.0978549316713615e-07,
"loss": 0.7447,
"step": 929
},
{
"epoch": 2.822190611664296,
"grad_norm": 2.100362539291382,
"learning_rate": 2.0263819378466888e-07,
"loss": 0.7393,
"step": 930
},
{
"epoch": 2.8252252252252252,
"grad_norm": 2.2228076457977295,
"learning_rate": 1.9561352038673264e-07,
"loss": 0.7501,
"step": 931
},
{
"epoch": 2.8282598387861544,
"grad_norm": 2.2403063774108887,
"learning_rate": 1.8871156089572018e-07,
"loss": 0.7508,
"step": 932
},
{
"epoch": 2.831294452347084,
"grad_norm": 2.1292290687561035,
"learning_rate": 1.8193240169810943e-07,
"loss": 0.7362,
"step": 933
},
{
"epoch": 2.834329065908013,
"grad_norm": 2.029907703399658,
"learning_rate": 1.752761276433801e-07,
"loss": 0.7504,
"step": 934
},
{
"epoch": 2.8373636794689427,
"grad_norm": 2.175046682357788,
"learning_rate": 1.6874282204295765e-07,
"loss": 0.7437,
"step": 935
},
{
"epoch": 2.840398293029872,
"grad_norm": 2.150054693222046,
"learning_rate": 1.623325666691644e-07,
"loss": 0.7388,
"step": 936
},
{
"epoch": 2.8434329065908015,
"grad_norm": 2.2034387588500977,
"learning_rate": 1.5604544175419901e-07,
"loss": 0.7451,
"step": 937
},
{
"epoch": 2.8464675201517307,
"grad_norm": 2.1042325496673584,
"learning_rate": 1.4988152598913063e-07,
"loss": 0.7535,
"step": 938
},
{
"epoch": 2.84950213371266,
"grad_norm": 2.1048696041107178,
"learning_rate": 1.4384089652291544e-07,
"loss": 0.745,
"step": 939
},
{
"epoch": 2.8525367472735894,
"grad_norm": 2.3105156421661377,
"learning_rate": 1.3792362896143164e-07,
"loss": 0.7233,
"step": 940
},
{
"epoch": 2.8555713608345186,
"grad_norm": 2.1417715549468994,
"learning_rate": 1.3212979736653142e-07,
"loss": 0.7467,
"step": 941
},
{
"epoch": 2.858605974395448,
"grad_norm": 2.1809537410736084,
"learning_rate": 1.2645947425511397e-07,
"loss": 0.7432,
"step": 942
},
{
"epoch": 2.8616405879563773,
"grad_norm": 2.2112855911254883,
"learning_rate": 1.209127305982205e-07,
"loss": 0.7487,
"step": 943
},
{
"epoch": 2.864675201517307,
"grad_norm": 2.059002161026001,
"learning_rate": 1.1548963582013961e-07,
"loss": 0.734,
"step": 944
},
{
"epoch": 2.867709815078236,
"grad_norm": 2.04377818107605,
"learning_rate": 1.1019025779754666e-07,
"loss": 0.7213,
"step": 945
},
{
"epoch": 2.8707444286391652,
"grad_norm": 2.0912530422210693,
"learning_rate": 1.0501466285865124e-07,
"loss": 0.7548,
"step": 946
},
{
"epoch": 2.873779042200095,
"grad_norm": 2.2096190452575684,
"learning_rate": 9.996291578236228e-08,
"loss": 0.7246,
"step": 947
},
{
"epoch": 2.8768136557610244,
"grad_norm": 2.284653663635254,
"learning_rate": 9.503507979748305e-08,
"loss": 0.7339,
"step": 948
},
{
"epoch": 2.8798482693219536,
"grad_norm": 2.097752332687378,
"learning_rate": 9.023121658191636e-08,
"loss": 0.7303,
"step": 949
},
{
"epoch": 2.8828828828828827,
"grad_norm": 2.171391010284424,
"learning_rate": 8.555138626189619e-08,
"loss": 0.7503,
"step": 950
},
{
"epoch": 2.8828828828828827,
"eval_loss": NaN,
"eval_runtime": 204.3654,
"eval_samples_per_second": 20.346,
"eval_steps_per_second": 5.089,
"step": 950
},
{
"epoch": 2.8859174964438123,
"grad_norm": 2.1366286277770996,
"learning_rate": 8.099564741123167e-08,
"loss": 0.7241,
"step": 951
},
{
"epoch": 2.8889521100047415,
"grad_norm": 2.0811874866485596,
"learning_rate": 7.656405705057434e-08,
"loss": 0.7504,
"step": 952
},
{
"epoch": 2.8919867235656707,
"grad_norm": 2.080226421356201,
"learning_rate": 7.225667064670761e-08,
"loss": 0.7355,
"step": 953
},
{
"epoch": 2.8950213371266003,
"grad_norm": 2.200331211090088,
"learning_rate": 6.807354211184613e-08,
"loss": 0.7427,
"step": 954
},
{
"epoch": 2.89805595068753,
"grad_norm": 2.038591146469116,
"learning_rate": 6.401472380297091e-08,
"loss": 0.7379,
"step": 955
},
{
"epoch": 2.901090564248459,
"grad_norm": 2.1141886711120605,
"learning_rate": 6.008026652116306e-08,
"loss": 0.7376,
"step": 956
},
{
"epoch": 2.904125177809388,
"grad_norm": 2.0803427696228027,
"learning_rate": 5.6270219510975445e-08,
"loss": 0.7424,
"step": 957
},
{
"epoch": 2.9071597913703178,
"grad_norm": 2.0900638103485107,
"learning_rate": 5.258463045981432e-08,
"loss": 0.7243,
"step": 958
},
{
"epoch": 2.910194404931247,
"grad_norm": 2.133312463760376,
"learning_rate": 4.902354549733979e-08,
"loss": 0.7257,
"step": 959
},
{
"epoch": 2.9132290184921765,
"grad_norm": 2.2082343101501465,
"learning_rate": 4.5587009194894005e-08,
"loss": 0.7346,
"step": 960
},
{
"epoch": 2.9162636320531057,
"grad_norm": 2.0872416496276855,
"learning_rate": 4.227506456493835e-08,
"loss": 0.7437,
"step": 961
},
{
"epoch": 2.9192982456140353,
"grad_norm": 2.136866807937622,
"learning_rate": 3.908775306051604e-08,
"loss": 0.7362,
"step": 962
},
{
"epoch": 2.9223328591749644,
"grad_norm": 2.125523805618286,
"learning_rate": 3.602511457473479e-08,
"loss": 0.7298,
"step": 963
},
{
"epoch": 2.9253674727358936,
"grad_norm": 2.055690050125122,
"learning_rate": 3.3087187440268284e-08,
"loss": 0.7285,
"step": 964
},
{
"epoch": 2.928402086296823,
"grad_norm": 2.26811146736145,
"learning_rate": 3.027400842887218e-08,
"loss": 0.732,
"step": 965
},
{
"epoch": 2.9314366998577523,
"grad_norm": 2.0984137058258057,
"learning_rate": 2.758561275092886e-08,
"loss": 0.7238,
"step": 966
},
{
"epoch": 2.934471313418682,
"grad_norm": 2.117258310317993,
"learning_rate": 2.5022034055003363e-08,
"loss": 0.7604,
"step": 967
},
{
"epoch": 2.937505926979611,
"grad_norm": 2.135441541671753,
"learning_rate": 2.2583304427421516e-08,
"loss": 0.7358,
"step": 968
},
{
"epoch": 2.9405405405405407,
"grad_norm": 2.1416280269622803,
"learning_rate": 2.0269454391874665e-08,
"loss": 0.7436,
"step": 969
},
{
"epoch": 2.94357515410147,
"grad_norm": 2.1733927726745605,
"learning_rate": 1.80805129090289e-08,
"loss": 0.7308,
"step": 970
},
{
"epoch": 2.946609767662399,
"grad_norm": 2.1229727268218994,
"learning_rate": 1.6016507376169776e-08,
"loss": 0.7577,
"step": 971
},
{
"epoch": 2.9496443812233286,
"grad_norm": 2.283834218978882,
"learning_rate": 1.4077463626852584e-08,
"loss": 0.748,
"step": 972
},
{
"epoch": 2.952678994784258,
"grad_norm": 2.205937623977661,
"learning_rate": 1.2263405930585947e-08,
"loss": 0.7335,
"step": 973
},
{
"epoch": 2.9557136083451874,
"grad_norm": 2.111551284790039,
"learning_rate": 1.0574356992525403e-08,
"loss": 0.7295,
"step": 974
},
{
"epoch": 2.9587482219061165,
"grad_norm": 2.058469295501709,
"learning_rate": 9.010337953185843e-09,
"loss": 0.7547,
"step": 975
},
{
"epoch": 2.961782835467046,
"grad_norm": 2.1459245681762695,
"learning_rate": 7.571368388181732e-09,
"loss": 0.7368,
"step": 976
},
{
"epoch": 2.9648174490279753,
"grad_norm": 2.1419780254364014,
"learning_rate": 6.257466307980631e-09,
"loss": 0.7364,
"step": 977
},
{
"epoch": 2.9678520625889044,
"grad_norm": 2.1368930339813232,
"learning_rate": 5.068648157675604e-09,
"loss": 0.7243,
"step": 978
},
{
"epoch": 2.970886676149834,
"grad_norm": 2.072345733642578,
"learning_rate": 4.00492881678427e-09,
"loss": 0.7229,
"step": 979
},
{
"epoch": 2.9739212897107636,
"grad_norm": 2.173828363418579,
"learning_rate": 3.0663215990534013e-09,
"loss": 0.711,
"step": 980
},
{
"epoch": 2.976955903271693,
"grad_norm": 2.117293357849121,
"learning_rate": 2.2528382523057115e-09,
"loss": 0.7648,
"step": 981
},
{
"epoch": 2.979990516832622,
"grad_norm": 2.126967191696167,
"learning_rate": 1.564488958279986e-09,
"loss": 0.7359,
"step": 982
},
{
"epoch": 2.9830251303935515,
"grad_norm": 2.086338520050049,
"learning_rate": 1.0012823325111776e-09,
"loss": 0.7575,
"step": 983
},
{
"epoch": 2.9860597439544807,
"grad_norm": 2.0174200534820557,
"learning_rate": 5.632254242204926e-10,
"loss": 0.7324,
"step": 984
},
{
"epoch": 2.9890943575154103,
"grad_norm": 2.0716025829315186,
"learning_rate": 2.503237162254646e-10,
"loss": 0.7482,
"step": 985
},
{
"epoch": 2.9921289710763395,
"grad_norm": 2.0842347145080566,
"learning_rate": 6.258112487667056e-11,
"loss": 0.7334,
"step": 986
},
{
"epoch": 2.995163584637269,
"grad_norm": 2.1061208248138428,
"learning_rate": 0.0,
"loss": 0.7299,
"step": 987
},
{
"epoch": 2.995163584637269,
"step": 987,
"total_flos": 4.415483185831556e+19,
"train_loss": 0.7698779804849093,
"train_runtime": 131039.1076,
"train_samples_per_second": 3.863,
"train_steps_per_second": 0.008
}
],
"logging_steps": 1.0,
"max_steps": 987,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4.415483185831556e+19,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}