lesso's picture
Training in progress, epoch 0, checkpoint
4f5fc32 verified
raw
history blame
176 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.10707784559374665,
"eval_steps": 500,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00010707784559374665,
"grad_norm": 0.4914730489253998,
"learning_rate": 0.00019999950652018584,
"loss": 1.0891,
"step": 1
},
{
"epoch": 0.0002141556911874933,
"grad_norm": 0.3680574297904968,
"learning_rate": 0.0001999980260856137,
"loss": 1.0538,
"step": 2
},
{
"epoch": 0.00032123353678124,
"grad_norm": 0.39417698979377747,
"learning_rate": 0.000199995558710895,
"loss": 1.0477,
"step": 3
},
{
"epoch": 0.0004283113823749866,
"grad_norm": 0.4732518792152405,
"learning_rate": 0.00019999210442038162,
"loss": 0.9224,
"step": 4
},
{
"epoch": 0.0005353892279687332,
"grad_norm": 0.41816651821136475,
"learning_rate": 0.00019998766324816607,
"loss": 0.8018,
"step": 5
},
{
"epoch": 0.00064246707356248,
"grad_norm": 0.45390111207962036,
"learning_rate": 0.0001999822352380809,
"loss": 0.7244,
"step": 6
},
{
"epoch": 0.0007495449191562266,
"grad_norm": 0.47834527492523193,
"learning_rate": 0.00019997582044369843,
"loss": 0.6486,
"step": 7
},
{
"epoch": 0.0008566227647499732,
"grad_norm": 0.4380565285682678,
"learning_rate": 0.00019996841892833,
"loss": 0.5942,
"step": 8
},
{
"epoch": 0.0009637006103437199,
"grad_norm": 0.5452256202697754,
"learning_rate": 0.00019996003076502565,
"loss": 0.5054,
"step": 9
},
{
"epoch": 0.0010707784559374665,
"grad_norm": 0.5409011244773865,
"learning_rate": 0.00019995065603657316,
"loss": 0.3999,
"step": 10
},
{
"epoch": 0.001177856301531213,
"grad_norm": 0.44740885496139526,
"learning_rate": 0.0001999402948354973,
"loss": 0.5802,
"step": 11
},
{
"epoch": 0.00128493414712496,
"grad_norm": 0.44699716567993164,
"learning_rate": 0.00019992894726405893,
"loss": 0.4222,
"step": 12
},
{
"epoch": 0.0013920119927187066,
"grad_norm": 0.4242172837257385,
"learning_rate": 0.000199916613434254,
"loss": 0.4839,
"step": 13
},
{
"epoch": 0.0014990898383124532,
"grad_norm": 0.3534027934074402,
"learning_rate": 0.0001999032934678125,
"loss": 0.3729,
"step": 14
},
{
"epoch": 0.0016061676839061998,
"grad_norm": 0.3338993191719055,
"learning_rate": 0.00019988898749619702,
"loss": 0.3565,
"step": 15
},
{
"epoch": 0.0017132455294999465,
"grad_norm": 0.5204335451126099,
"learning_rate": 0.00019987369566060176,
"loss": 0.3462,
"step": 16
},
{
"epoch": 0.001820323375093693,
"grad_norm": 0.3927958905696869,
"learning_rate": 0.00019985741811195097,
"loss": 0.455,
"step": 17
},
{
"epoch": 0.0019274012206874397,
"grad_norm": 0.28747662901878357,
"learning_rate": 0.00019984015501089752,
"loss": 0.3915,
"step": 18
},
{
"epoch": 0.0020344790662811863,
"grad_norm": 0.34313827753067017,
"learning_rate": 0.0001998219065278212,
"loss": 0.4298,
"step": 19
},
{
"epoch": 0.002141556911874933,
"grad_norm": 0.3621661067008972,
"learning_rate": 0.00019980267284282717,
"loss": 0.3907,
"step": 20
},
{
"epoch": 0.0022486347574686796,
"grad_norm": 0.3450513780117035,
"learning_rate": 0.00019978245414574417,
"loss": 0.3854,
"step": 21
},
{
"epoch": 0.002355712603062426,
"grad_norm": 0.3595326244831085,
"learning_rate": 0.00019976125063612252,
"loss": 0.3459,
"step": 22
},
{
"epoch": 0.002462790448656173,
"grad_norm": 0.7127284407615662,
"learning_rate": 0.00019973906252323238,
"loss": 0.2965,
"step": 23
},
{
"epoch": 0.00256986829424992,
"grad_norm": 0.40753620862960815,
"learning_rate": 0.0001997158900260614,
"loss": 0.4976,
"step": 24
},
{
"epoch": 0.0026769461398436665,
"grad_norm": 0.25613075494766235,
"learning_rate": 0.0001996917333733128,
"loss": 0.3591,
"step": 25
},
{
"epoch": 0.002784023985437413,
"grad_norm": 0.42440053820610046,
"learning_rate": 0.00019966659280340297,
"loss": 0.3673,
"step": 26
},
{
"epoch": 0.0028911018310311598,
"grad_norm": 0.2876407206058502,
"learning_rate": 0.00019964046856445924,
"loss": 0.2941,
"step": 27
},
{
"epoch": 0.0029981796766249064,
"grad_norm": 0.2591134011745453,
"learning_rate": 0.00019961336091431727,
"loss": 0.297,
"step": 28
},
{
"epoch": 0.003105257522218653,
"grad_norm": 0.25432032346725464,
"learning_rate": 0.00019958527012051857,
"loss": 0.2986,
"step": 29
},
{
"epoch": 0.0032123353678123997,
"grad_norm": 0.29807740449905396,
"learning_rate": 0.00019955619646030802,
"loss": 0.4165,
"step": 30
},
{
"epoch": 0.0033194132134061463,
"grad_norm": 0.22817127406597137,
"learning_rate": 0.00019952614022063084,
"loss": 0.3591,
"step": 31
},
{
"epoch": 0.003426491058999893,
"grad_norm": 0.2147909253835678,
"learning_rate": 0.00019949510169813003,
"loss": 0.3804,
"step": 32
},
{
"epoch": 0.0035335689045936395,
"grad_norm": 0.26951637864112854,
"learning_rate": 0.00019946308119914323,
"loss": 0.3321,
"step": 33
},
{
"epoch": 0.003640646750187386,
"grad_norm": 0.2519189119338989,
"learning_rate": 0.0001994300790396999,
"loss": 0.3917,
"step": 34
},
{
"epoch": 0.0037477245957811328,
"grad_norm": 0.2383698970079422,
"learning_rate": 0.000199396095545518,
"loss": 0.3692,
"step": 35
},
{
"epoch": 0.0038548024413748794,
"grad_norm": 0.1950429528951645,
"learning_rate": 0.00019936113105200085,
"loss": 0.2132,
"step": 36
},
{
"epoch": 0.003961880286968626,
"grad_norm": 0.24658283591270447,
"learning_rate": 0.00019932518590423394,
"loss": 0.3589,
"step": 37
},
{
"epoch": 0.004068958132562373,
"grad_norm": 0.21331077814102173,
"learning_rate": 0.00019928826045698136,
"loss": 0.3155,
"step": 38
},
{
"epoch": 0.004176035978156119,
"grad_norm": 0.25296568870544434,
"learning_rate": 0.0001992503550746824,
"loss": 0.3137,
"step": 39
},
{
"epoch": 0.004283113823749866,
"grad_norm": 0.23738770186901093,
"learning_rate": 0.0001992114701314478,
"loss": 0.3445,
"step": 40
},
{
"epoch": 0.0043901916693436125,
"grad_norm": 0.23128816485404968,
"learning_rate": 0.0001991716060110563,
"loss": 0.3027,
"step": 41
},
{
"epoch": 0.004497269514937359,
"grad_norm": 0.21762651205062866,
"learning_rate": 0.00019913076310695068,
"loss": 0.3852,
"step": 42
},
{
"epoch": 0.004604347360531106,
"grad_norm": 0.2557142972946167,
"learning_rate": 0.00019908894182223388,
"loss": 0.3346,
"step": 43
},
{
"epoch": 0.004711425206124852,
"grad_norm": 0.21737657487392426,
"learning_rate": 0.00019904614256966512,
"loss": 0.3461,
"step": 44
},
{
"epoch": 0.004818503051718599,
"grad_norm": 0.21647535264492035,
"learning_rate": 0.00019900236577165576,
"loss": 0.2972,
"step": 45
},
{
"epoch": 0.004925580897312346,
"grad_norm": 0.24528199434280396,
"learning_rate": 0.0001989576118602651,
"loss": 0.349,
"step": 46
},
{
"epoch": 0.005032658742906092,
"grad_norm": 0.22419703006744385,
"learning_rate": 0.00019891188127719618,
"loss": 0.3198,
"step": 47
},
{
"epoch": 0.00513973658849984,
"grad_norm": 0.18803644180297852,
"learning_rate": 0.0001988651744737914,
"loss": 0.2826,
"step": 48
},
{
"epoch": 0.005246814434093586,
"grad_norm": 0.23534414172172546,
"learning_rate": 0.00019881749191102808,
"loss": 0.3813,
"step": 49
},
{
"epoch": 0.005353892279687333,
"grad_norm": 0.1908506602048874,
"learning_rate": 0.00019876883405951377,
"loss": 0.3349,
"step": 50
},
{
"epoch": 0.00546097012528108,
"grad_norm": 0.22902971506118774,
"learning_rate": 0.00019871920139948192,
"loss": 0.4143,
"step": 51
},
{
"epoch": 0.005568047970874826,
"grad_norm": 0.19370073080062866,
"learning_rate": 0.0001986685944207868,
"loss": 0.3182,
"step": 52
},
{
"epoch": 0.005675125816468573,
"grad_norm": 0.22496962547302246,
"learning_rate": 0.0001986170136228989,
"loss": 0.3622,
"step": 53
},
{
"epoch": 0.0057822036620623195,
"grad_norm": 0.19907711446285248,
"learning_rate": 0.00019856445951489982,
"loss": 0.3328,
"step": 54
},
{
"epoch": 0.005889281507656066,
"grad_norm": 0.28724735975265503,
"learning_rate": 0.0001985109326154774,
"loss": 0.4235,
"step": 55
},
{
"epoch": 0.005996359353249813,
"grad_norm": 0.2073618769645691,
"learning_rate": 0.00019845643345292054,
"loss": 0.3651,
"step": 56
},
{
"epoch": 0.006103437198843559,
"grad_norm": 0.3525002896785736,
"learning_rate": 0.00019840096256511398,
"loss": 0.4431,
"step": 57
},
{
"epoch": 0.006210515044437306,
"grad_norm": 0.2288159877061844,
"learning_rate": 0.00019834452049953297,
"loss": 0.3314,
"step": 58
},
{
"epoch": 0.006317592890031053,
"grad_norm": 0.27389129996299744,
"learning_rate": 0.00019828710781323792,
"loss": 0.4399,
"step": 59
},
{
"epoch": 0.006424670735624799,
"grad_norm": 0.37030190229415894,
"learning_rate": 0.0001982287250728689,
"loss": 0.4215,
"step": 60
},
{
"epoch": 0.006531748581218546,
"grad_norm": 0.19040270149707794,
"learning_rate": 0.0001981693728546399,
"loss": 0.3156,
"step": 61
},
{
"epoch": 0.0066388264268122926,
"grad_norm": 0.23044194281101227,
"learning_rate": 0.0001981090517443334,
"loss": 0.3071,
"step": 62
},
{
"epoch": 0.006745904272406039,
"grad_norm": 0.2296312153339386,
"learning_rate": 0.00019804776233729444,
"loss": 0.3353,
"step": 63
},
{
"epoch": 0.006852982117999786,
"grad_norm": 0.2212522029876709,
"learning_rate": 0.0001979855052384247,
"loss": 0.3534,
"step": 64
},
{
"epoch": 0.0069600599635935324,
"grad_norm": 0.23206260800361633,
"learning_rate": 0.00019792228106217658,
"loss": 0.3441,
"step": 65
},
{
"epoch": 0.007067137809187279,
"grad_norm": 0.19569985568523407,
"learning_rate": 0.00019785809043254722,
"loss": 0.335,
"step": 66
},
{
"epoch": 0.007174215654781026,
"grad_norm": 0.23851607739925385,
"learning_rate": 0.0001977929339830722,
"loss": 0.487,
"step": 67
},
{
"epoch": 0.007281293500374772,
"grad_norm": 0.19546416401863098,
"learning_rate": 0.00019772681235681936,
"loss": 0.44,
"step": 68
},
{
"epoch": 0.007388371345968519,
"grad_norm": 0.20493975281715393,
"learning_rate": 0.00019765972620638248,
"loss": 0.3202,
"step": 69
},
{
"epoch": 0.0074954491915622656,
"grad_norm": 0.1694723516702652,
"learning_rate": 0.00019759167619387476,
"loss": 0.3613,
"step": 70
},
{
"epoch": 0.007602527037156012,
"grad_norm": 0.29382210969924927,
"learning_rate": 0.00019752266299092236,
"loss": 0.3509,
"step": 71
},
{
"epoch": 0.007709604882749759,
"grad_norm": 0.17875567078590393,
"learning_rate": 0.00019745268727865774,
"loss": 0.332,
"step": 72
},
{
"epoch": 0.007816682728343505,
"grad_norm": 0.1698933243751526,
"learning_rate": 0.0001973817497477129,
"loss": 0.2699,
"step": 73
},
{
"epoch": 0.007923760573937252,
"grad_norm": 0.17154669761657715,
"learning_rate": 0.00019730985109821266,
"loss": 0.3798,
"step": 74
},
{
"epoch": 0.008030838419530999,
"grad_norm": 0.19141319394111633,
"learning_rate": 0.00019723699203976766,
"loss": 0.3966,
"step": 75
},
{
"epoch": 0.008137916265124745,
"grad_norm": 0.1600220948457718,
"learning_rate": 0.0001971631732914674,
"loss": 0.2993,
"step": 76
},
{
"epoch": 0.008244994110718492,
"grad_norm": 0.1882285475730896,
"learning_rate": 0.0001970883955818731,
"loss": 0.2683,
"step": 77
},
{
"epoch": 0.008352071956312239,
"grad_norm": 0.17301051318645477,
"learning_rate": 0.0001970126596490106,
"loss": 0.2705,
"step": 78
},
{
"epoch": 0.008459149801905985,
"grad_norm": 0.21556293964385986,
"learning_rate": 0.00019693596624036292,
"loss": 0.4031,
"step": 79
},
{
"epoch": 0.008566227647499732,
"grad_norm": 0.1741161048412323,
"learning_rate": 0.0001968583161128631,
"loss": 0.3607,
"step": 80
},
{
"epoch": 0.008673305493093478,
"grad_norm": 0.25049349665641785,
"learning_rate": 0.00019677971003288655,
"loss": 0.5397,
"step": 81
},
{
"epoch": 0.008780383338687225,
"grad_norm": 0.16967053711414337,
"learning_rate": 0.00019670014877624353,
"loss": 0.3115,
"step": 82
},
{
"epoch": 0.008887461184280972,
"grad_norm": 0.26333314180374146,
"learning_rate": 0.00019661963312817148,
"loss": 0.3876,
"step": 83
},
{
"epoch": 0.008994539029874718,
"grad_norm": 0.19522804021835327,
"learning_rate": 0.0001965381638833274,
"loss": 0.3693,
"step": 84
},
{
"epoch": 0.009101616875468465,
"grad_norm": 0.17835591733455658,
"learning_rate": 0.00019645574184577982,
"loss": 0.3377,
"step": 85
},
{
"epoch": 0.009208694721062212,
"grad_norm": 0.19591933488845825,
"learning_rate": 0.000196372367829001,
"loss": 0.3446,
"step": 86
},
{
"epoch": 0.009315772566655958,
"grad_norm": 0.17983509600162506,
"learning_rate": 0.00019628804265585877,
"loss": 0.3993,
"step": 87
},
{
"epoch": 0.009422850412249705,
"grad_norm": 0.1638455092906952,
"learning_rate": 0.0001962027671586086,
"loss": 0.2946,
"step": 88
},
{
"epoch": 0.009529928257843451,
"grad_norm": 0.17286810278892517,
"learning_rate": 0.0001961165421788852,
"loss": 0.3427,
"step": 89
},
{
"epoch": 0.009637006103437198,
"grad_norm": 0.18723516166210175,
"learning_rate": 0.0001960293685676943,
"loss": 0.2939,
"step": 90
},
{
"epoch": 0.009744083949030945,
"grad_norm": 0.2359951138496399,
"learning_rate": 0.0001959412471854043,
"loss": 0.4401,
"step": 91
},
{
"epoch": 0.009851161794624691,
"grad_norm": 0.1666853427886963,
"learning_rate": 0.0001958521789017376,
"loss": 0.3019,
"step": 92
},
{
"epoch": 0.009958239640218438,
"grad_norm": 0.20216096937656403,
"learning_rate": 0.00019576216459576222,
"loss": 0.3507,
"step": 93
},
{
"epoch": 0.010065317485812185,
"grad_norm": 0.1928347647190094,
"learning_rate": 0.00019567120515588308,
"loss": 0.4894,
"step": 94
},
{
"epoch": 0.010172395331405933,
"grad_norm": 0.2161427140235901,
"learning_rate": 0.00019557930147983302,
"loss": 0.4311,
"step": 95
},
{
"epoch": 0.01027947317699968,
"grad_norm": 0.1798643171787262,
"learning_rate": 0.00019548645447466431,
"loss": 0.2975,
"step": 96
},
{
"epoch": 0.010386551022593426,
"grad_norm": 0.21056245267391205,
"learning_rate": 0.00019539266505673938,
"loss": 0.3106,
"step": 97
},
{
"epoch": 0.010493628868187173,
"grad_norm": 0.22675903141498566,
"learning_rate": 0.00019529793415172192,
"loss": 0.4743,
"step": 98
},
{
"epoch": 0.01060070671378092,
"grad_norm": 0.19647933542728424,
"learning_rate": 0.00019520226269456768,
"loss": 0.319,
"step": 99
},
{
"epoch": 0.010707784559374666,
"grad_norm": 0.15131019055843353,
"learning_rate": 0.00019510565162951537,
"loss": 0.2944,
"step": 100
},
{
"epoch": 0.010814862404968413,
"grad_norm": 0.18563662469387054,
"learning_rate": 0.00019500810191007718,
"loss": 0.3675,
"step": 101
},
{
"epoch": 0.01092194025056216,
"grad_norm": 0.203588604927063,
"learning_rate": 0.00019490961449902946,
"loss": 0.3694,
"step": 102
},
{
"epoch": 0.011029018096155906,
"grad_norm": 0.22488757967948914,
"learning_rate": 0.0001948101903684032,
"loss": 0.4804,
"step": 103
},
{
"epoch": 0.011136095941749653,
"grad_norm": 0.17686226963996887,
"learning_rate": 0.00019470983049947444,
"loss": 0.3596,
"step": 104
},
{
"epoch": 0.0112431737873434,
"grad_norm": 0.15883031487464905,
"learning_rate": 0.00019460853588275454,
"loss": 0.3952,
"step": 105
},
{
"epoch": 0.011350251632937146,
"grad_norm": 0.1682022213935852,
"learning_rate": 0.00019450630751798048,
"loss": 0.2541,
"step": 106
},
{
"epoch": 0.011457329478530892,
"grad_norm": 0.22802792489528656,
"learning_rate": 0.000194403146414105,
"loss": 0.3886,
"step": 107
},
{
"epoch": 0.011564407324124639,
"grad_norm": 0.17755688726902008,
"learning_rate": 0.00019429905358928646,
"loss": 0.3412,
"step": 108
},
{
"epoch": 0.011671485169718386,
"grad_norm": 0.20976732671260834,
"learning_rate": 0.00019419403007087907,
"loss": 0.3449,
"step": 109
},
{
"epoch": 0.011778563015312132,
"grad_norm": 0.1593596190214157,
"learning_rate": 0.00019408807689542257,
"loss": 0.3573,
"step": 110
},
{
"epoch": 0.011885640860905879,
"grad_norm": 0.20216397941112518,
"learning_rate": 0.00019398119510863197,
"loss": 0.4127,
"step": 111
},
{
"epoch": 0.011992718706499626,
"grad_norm": 0.1636137217283249,
"learning_rate": 0.00019387338576538744,
"loss": 0.2631,
"step": 112
},
{
"epoch": 0.012099796552093372,
"grad_norm": 0.18087638914585114,
"learning_rate": 0.00019376464992972356,
"loss": 0.3848,
"step": 113
},
{
"epoch": 0.012206874397687119,
"grad_norm": 0.22892820835113525,
"learning_rate": 0.00019365498867481923,
"loss": 0.4325,
"step": 114
},
{
"epoch": 0.012313952243280865,
"grad_norm": 0.16127265989780426,
"learning_rate": 0.00019354440308298675,
"loss": 0.2838,
"step": 115
},
{
"epoch": 0.012421030088874612,
"grad_norm": 0.17154908180236816,
"learning_rate": 0.00019343289424566122,
"loss": 0.4173,
"step": 116
},
{
"epoch": 0.012528107934468359,
"grad_norm": 0.13716156780719757,
"learning_rate": 0.00019332046326338986,
"loss": 0.2642,
"step": 117
},
{
"epoch": 0.012635185780062105,
"grad_norm": 0.16628596186637878,
"learning_rate": 0.0001932071112458211,
"loss": 0.3342,
"step": 118
},
{
"epoch": 0.012742263625655852,
"grad_norm": 0.17461515963077545,
"learning_rate": 0.00019309283931169356,
"loss": 0.3829,
"step": 119
},
{
"epoch": 0.012849341471249599,
"grad_norm": 0.20186425745487213,
"learning_rate": 0.00019297764858882514,
"loss": 0.4213,
"step": 120
},
{
"epoch": 0.012956419316843345,
"grad_norm": 0.18099796772003174,
"learning_rate": 0.00019286154021410173,
"loss": 0.3205,
"step": 121
},
{
"epoch": 0.013063497162437092,
"grad_norm": 0.19527661800384521,
"learning_rate": 0.00019274451533346615,
"loss": 0.4162,
"step": 122
},
{
"epoch": 0.013170575008030838,
"grad_norm": 0.16075778007507324,
"learning_rate": 0.00019262657510190666,
"loss": 0.2801,
"step": 123
},
{
"epoch": 0.013277652853624585,
"grad_norm": 0.16372549533843994,
"learning_rate": 0.0001925077206834458,
"loss": 0.3641,
"step": 124
},
{
"epoch": 0.013384730699218332,
"grad_norm": 0.18662606179714203,
"learning_rate": 0.0001923879532511287,
"loss": 0.3799,
"step": 125
},
{
"epoch": 0.013491808544812078,
"grad_norm": 0.1677158623933792,
"learning_rate": 0.0001922672739870115,
"loss": 0.2898,
"step": 126
},
{
"epoch": 0.013598886390405825,
"grad_norm": 0.19273678958415985,
"learning_rate": 0.00019214568408214985,
"loss": 0.3585,
"step": 127
},
{
"epoch": 0.013705964235999572,
"grad_norm": 0.20474790036678314,
"learning_rate": 0.00019202318473658705,
"loss": 0.4194,
"step": 128
},
{
"epoch": 0.013813042081593318,
"grad_norm": 0.19291414320468903,
"learning_rate": 0.00019189977715934213,
"loss": 0.4429,
"step": 129
},
{
"epoch": 0.013920119927187065,
"grad_norm": 0.19790148735046387,
"learning_rate": 0.00019177546256839812,
"loss": 0.3079,
"step": 130
},
{
"epoch": 0.014027197772780811,
"grad_norm": 0.1970977485179901,
"learning_rate": 0.0001916502421906898,
"loss": 0.3948,
"step": 131
},
{
"epoch": 0.014134275618374558,
"grad_norm": 0.18109802901744843,
"learning_rate": 0.00019152411726209176,
"loss": 0.2952,
"step": 132
},
{
"epoch": 0.014241353463968305,
"grad_norm": 0.16982200741767883,
"learning_rate": 0.00019139708902740613,
"loss": 0.3422,
"step": 133
},
{
"epoch": 0.014348431309562051,
"grad_norm": 0.17093245685100555,
"learning_rate": 0.0001912691587403503,
"loss": 0.3613,
"step": 134
},
{
"epoch": 0.014455509155155798,
"grad_norm": 0.15467596054077148,
"learning_rate": 0.00019114032766354453,
"loss": 0.2985,
"step": 135
},
{
"epoch": 0.014562587000749545,
"grad_norm": 0.20507284998893738,
"learning_rate": 0.00019101059706849957,
"loss": 0.3405,
"step": 136
},
{
"epoch": 0.014669664846343291,
"grad_norm": 0.17259500920772552,
"learning_rate": 0.00019087996823560402,
"loss": 0.4002,
"step": 137
},
{
"epoch": 0.014776742691937038,
"grad_norm": 0.1541275680065155,
"learning_rate": 0.0001907484424541117,
"loss": 0.2888,
"step": 138
},
{
"epoch": 0.014883820537530785,
"grad_norm": 0.19659638404846191,
"learning_rate": 0.00019061602102212898,
"loss": 0.4393,
"step": 139
},
{
"epoch": 0.014990898383124531,
"grad_norm": 0.17995648086071014,
"learning_rate": 0.00019048270524660196,
"loss": 0.333,
"step": 140
},
{
"epoch": 0.015097976228718278,
"grad_norm": 0.22955287992954254,
"learning_rate": 0.0001903484964433035,
"loss": 0.2958,
"step": 141
},
{
"epoch": 0.015205054074312024,
"grad_norm": 0.164543017745018,
"learning_rate": 0.00019021339593682028,
"loss": 0.3002,
"step": 142
},
{
"epoch": 0.015312131919905771,
"grad_norm": 0.15457768738269806,
"learning_rate": 0.00019007740506053983,
"loss": 0.3089,
"step": 143
},
{
"epoch": 0.015419209765499518,
"grad_norm": 0.17177622020244598,
"learning_rate": 0.0001899405251566371,
"loss": 0.3263,
"step": 144
},
{
"epoch": 0.015526287611093264,
"grad_norm": 0.1843390315771103,
"learning_rate": 0.00018980275757606157,
"loss": 0.3417,
"step": 145
},
{
"epoch": 0.01563336545668701,
"grad_norm": 0.15636736154556274,
"learning_rate": 0.00018966410367852362,
"loss": 0.2202,
"step": 146
},
{
"epoch": 0.01574044330228076,
"grad_norm": 0.1969316303730011,
"learning_rate": 0.00018952456483248119,
"loss": 0.4284,
"step": 147
},
{
"epoch": 0.015847521147874504,
"grad_norm": 0.17580685019493103,
"learning_rate": 0.0001893841424151264,
"loss": 0.2754,
"step": 148
},
{
"epoch": 0.015954598993468252,
"grad_norm": 0.1819283366203308,
"learning_rate": 0.0001892428378123718,
"loss": 0.3462,
"step": 149
},
{
"epoch": 0.016061676839061997,
"grad_norm": 0.1740637719631195,
"learning_rate": 0.0001891006524188368,
"loss": 0.3225,
"step": 150
},
{
"epoch": 0.016168754684655746,
"grad_norm": 0.17211860418319702,
"learning_rate": 0.00018895758763783383,
"loss": 0.3688,
"step": 151
},
{
"epoch": 0.01627583253024949,
"grad_norm": 0.1924057900905609,
"learning_rate": 0.00018881364488135448,
"loss": 0.3207,
"step": 152
},
{
"epoch": 0.01638291037584324,
"grad_norm": 0.16835831105709076,
"learning_rate": 0.00018866882557005567,
"loss": 0.3377,
"step": 153
},
{
"epoch": 0.016489988221436984,
"grad_norm": 0.15114548802375793,
"learning_rate": 0.00018852313113324552,
"loss": 0.2549,
"step": 154
},
{
"epoch": 0.016597066067030732,
"grad_norm": 0.14319011569023132,
"learning_rate": 0.00018837656300886937,
"loss": 0.2206,
"step": 155
},
{
"epoch": 0.016704143912624477,
"grad_norm": 0.18031498789787292,
"learning_rate": 0.00018822912264349534,
"loss": 0.4739,
"step": 156
},
{
"epoch": 0.016811221758218226,
"grad_norm": 0.25163549184799194,
"learning_rate": 0.00018808081149230036,
"loss": 0.4744,
"step": 157
},
{
"epoch": 0.01691829960381197,
"grad_norm": 0.16316188871860504,
"learning_rate": 0.00018793163101905563,
"loss": 0.3651,
"step": 158
},
{
"epoch": 0.01702537744940572,
"grad_norm": 0.1717788577079773,
"learning_rate": 0.00018778158269611218,
"loss": 0.31,
"step": 159
},
{
"epoch": 0.017132455294999464,
"grad_norm": 0.1477271467447281,
"learning_rate": 0.00018763066800438636,
"loss": 0.2215,
"step": 160
},
{
"epoch": 0.017239533140593212,
"grad_norm": 0.15727677941322327,
"learning_rate": 0.0001874788884333453,
"loss": 0.3529,
"step": 161
},
{
"epoch": 0.017346610986186957,
"grad_norm": 0.1828835904598236,
"learning_rate": 0.00018732624548099204,
"loss": 0.3712,
"step": 162
},
{
"epoch": 0.017453688831780705,
"grad_norm": 0.18454089760780334,
"learning_rate": 0.0001871727406538509,
"loss": 0.3708,
"step": 163
},
{
"epoch": 0.01756076667737445,
"grad_norm": 0.26658540964126587,
"learning_rate": 0.0001870183754669526,
"loss": 0.3926,
"step": 164
},
{
"epoch": 0.0176678445229682,
"grad_norm": 0.16646800935268402,
"learning_rate": 0.00018686315144381913,
"loss": 0.2482,
"step": 165
},
{
"epoch": 0.017774922368561943,
"grad_norm": 0.2011224329471588,
"learning_rate": 0.000186707070116449,
"loss": 0.3693,
"step": 166
},
{
"epoch": 0.017882000214155692,
"grad_norm": 0.15221862494945526,
"learning_rate": 0.0001865501330253019,
"loss": 0.318,
"step": 167
},
{
"epoch": 0.017989078059749437,
"grad_norm": 0.22113820910453796,
"learning_rate": 0.00018639234171928353,
"loss": 0.3929,
"step": 168
},
{
"epoch": 0.018096155905343185,
"grad_norm": 0.18122068047523499,
"learning_rate": 0.0001862336977557304,
"loss": 0.4554,
"step": 169
},
{
"epoch": 0.01820323375093693,
"grad_norm": 0.16834978759288788,
"learning_rate": 0.0001860742027003944,
"loss": 0.3052,
"step": 170
},
{
"epoch": 0.01831031159653068,
"grad_norm": 0.18118542432785034,
"learning_rate": 0.00018591385812742725,
"loss": 0.2902,
"step": 171
},
{
"epoch": 0.018417389442124423,
"grad_norm": 0.18435969948768616,
"learning_rate": 0.00018575266561936523,
"loss": 0.3397,
"step": 172
},
{
"epoch": 0.01852446728771817,
"grad_norm": 0.18012931942939758,
"learning_rate": 0.00018559062676711332,
"loss": 0.2858,
"step": 173
},
{
"epoch": 0.018631545133311916,
"grad_norm": 0.2357625663280487,
"learning_rate": 0.0001854277431699295,
"loss": 0.3273,
"step": 174
},
{
"epoch": 0.018738622978905665,
"grad_norm": 0.19773603975772858,
"learning_rate": 0.00018526401643540922,
"loss": 0.3522,
"step": 175
},
{
"epoch": 0.01884570082449941,
"grad_norm": 0.1897646188735962,
"learning_rate": 0.00018509944817946922,
"loss": 0.422,
"step": 176
},
{
"epoch": 0.018952778670093158,
"grad_norm": 0.2005687654018402,
"learning_rate": 0.00018493404002633166,
"loss": 0.4571,
"step": 177
},
{
"epoch": 0.019059856515686903,
"grad_norm": 0.18125112354755402,
"learning_rate": 0.00018476779360850832,
"loss": 0.4359,
"step": 178
},
{
"epoch": 0.01916693436128065,
"grad_norm": 0.2057882696390152,
"learning_rate": 0.00018460071056678422,
"loss": 0.4458,
"step": 179
},
{
"epoch": 0.019274012206874396,
"grad_norm": 0.3153612017631531,
"learning_rate": 0.00018443279255020152,
"loss": 0.3945,
"step": 180
},
{
"epoch": 0.019381090052468145,
"grad_norm": 0.21111486852169037,
"learning_rate": 0.00018426404121604323,
"loss": 0.4705,
"step": 181
},
{
"epoch": 0.01948816789806189,
"grad_norm": 0.2238568216562271,
"learning_rate": 0.00018409445822981693,
"loss": 0.3776,
"step": 182
},
{
"epoch": 0.019595245743655638,
"grad_norm": 0.2252410352230072,
"learning_rate": 0.00018392404526523817,
"loss": 0.3556,
"step": 183
},
{
"epoch": 0.019702323589249383,
"grad_norm": 0.1595439463853836,
"learning_rate": 0.0001837528040042142,
"loss": 0.2908,
"step": 184
},
{
"epoch": 0.01980940143484313,
"grad_norm": 0.17755721509456635,
"learning_rate": 0.00018358073613682706,
"loss": 0.3642,
"step": 185
},
{
"epoch": 0.019916479280436876,
"grad_norm": 0.24803121387958527,
"learning_rate": 0.00018340784336131713,
"loss": 0.4246,
"step": 186
},
{
"epoch": 0.020023557126030624,
"grad_norm": 0.1807357370853424,
"learning_rate": 0.00018323412738406635,
"loss": 0.3031,
"step": 187
},
{
"epoch": 0.02013063497162437,
"grad_norm": 0.18305666744709015,
"learning_rate": 0.00018305958991958127,
"loss": 0.4081,
"step": 188
},
{
"epoch": 0.020237712817218118,
"grad_norm": 0.15840698778629303,
"learning_rate": 0.0001828842326904762,
"loss": 0.2594,
"step": 189
},
{
"epoch": 0.020344790662811866,
"grad_norm": 0.1579771190881729,
"learning_rate": 0.00018270805742745617,
"loss": 0.2373,
"step": 190
},
{
"epoch": 0.02045186850840561,
"grad_norm": 0.1966482698917389,
"learning_rate": 0.00018253106586929997,
"loss": 0.4255,
"step": 191
},
{
"epoch": 0.02055894635399936,
"grad_norm": 0.16462944447994232,
"learning_rate": 0.00018235325976284275,
"loss": 0.2783,
"step": 192
},
{
"epoch": 0.020666024199593104,
"grad_norm": 0.15610237419605255,
"learning_rate": 0.00018217464086295904,
"loss": 0.2818,
"step": 193
},
{
"epoch": 0.020773102045186852,
"grad_norm": 0.15715450048446655,
"learning_rate": 0.00018199521093254523,
"loss": 0.2853,
"step": 194
},
{
"epoch": 0.020880179890780597,
"grad_norm": 0.202264666557312,
"learning_rate": 0.00018181497174250236,
"loss": 0.3436,
"step": 195
},
{
"epoch": 0.020987257736374346,
"grad_norm": 0.21248966455459595,
"learning_rate": 0.00018163392507171842,
"loss": 0.3021,
"step": 196
},
{
"epoch": 0.02109433558196809,
"grad_norm": 0.14210785925388336,
"learning_rate": 0.00018145207270705096,
"loss": 0.246,
"step": 197
},
{
"epoch": 0.02120141342756184,
"grad_norm": 0.15923386812210083,
"learning_rate": 0.0001812694164433094,
"loss": 0.3208,
"step": 198
},
{
"epoch": 0.021308491273155584,
"grad_norm": 0.16584694385528564,
"learning_rate": 0.00018108595808323736,
"loss": 0.281,
"step": 199
},
{
"epoch": 0.021415569118749332,
"grad_norm": 0.1751406490802765,
"learning_rate": 0.00018090169943749476,
"loss": 0.3897,
"step": 200
},
{
"epoch": 0.021522646964343077,
"grad_norm": 0.1961715668439865,
"learning_rate": 0.00018071664232464002,
"loss": 0.4197,
"step": 201
},
{
"epoch": 0.021629724809936825,
"grad_norm": 0.1973417103290558,
"learning_rate": 0.0001805307885711122,
"loss": 0.3362,
"step": 202
},
{
"epoch": 0.02173680265553057,
"grad_norm": 0.16562364995479584,
"learning_rate": 0.00018034414001121278,
"loss": 0.4085,
"step": 203
},
{
"epoch": 0.02184388050112432,
"grad_norm": 0.1984337717294693,
"learning_rate": 0.00018015669848708767,
"loss": 0.2735,
"step": 204
},
{
"epoch": 0.021950958346718064,
"grad_norm": 0.18170015513896942,
"learning_rate": 0.00017996846584870908,
"loss": 0.3562,
"step": 205
},
{
"epoch": 0.022058036192311812,
"grad_norm": 0.1938287615776062,
"learning_rate": 0.0001797794439538571,
"loss": 0.3715,
"step": 206
},
{
"epoch": 0.022165114037905557,
"grad_norm": 0.16263467073440552,
"learning_rate": 0.0001795896346681016,
"loss": 0.3797,
"step": 207
},
{
"epoch": 0.022272191883499305,
"grad_norm": 0.1593952178955078,
"learning_rate": 0.00017939903986478355,
"loss": 0.2275,
"step": 208
},
{
"epoch": 0.02237926972909305,
"grad_norm": 0.19579128921031952,
"learning_rate": 0.00017920766142499672,
"loss": 0.318,
"step": 209
},
{
"epoch": 0.0224863475746868,
"grad_norm": 0.18321353197097778,
"learning_rate": 0.00017901550123756906,
"loss": 0.3672,
"step": 210
},
{
"epoch": 0.022593425420280543,
"grad_norm": 0.20865146815776825,
"learning_rate": 0.00017882256119904403,
"loss": 0.4745,
"step": 211
},
{
"epoch": 0.02270050326587429,
"grad_norm": 0.15021853148937225,
"learning_rate": 0.00017862884321366188,
"loss": 0.3397,
"step": 212
},
{
"epoch": 0.022807581111468037,
"grad_norm": 0.19131535291671753,
"learning_rate": 0.000178434349193341,
"loss": 0.3389,
"step": 213
},
{
"epoch": 0.022914658957061785,
"grad_norm": 0.1574268788099289,
"learning_rate": 0.0001782390810576588,
"loss": 0.3331,
"step": 214
},
{
"epoch": 0.02302173680265553,
"grad_norm": 0.18308749794960022,
"learning_rate": 0.000178043040733833,
"loss": 0.3105,
"step": 215
},
{
"epoch": 0.023128814648249278,
"grad_norm": 0.2218441218137741,
"learning_rate": 0.00017784623015670238,
"loss": 0.4791,
"step": 216
},
{
"epoch": 0.023235892493843023,
"grad_norm": 0.18416514992713928,
"learning_rate": 0.00017764865126870786,
"loss": 0.3551,
"step": 217
},
{
"epoch": 0.02334297033943677,
"grad_norm": 0.15702709555625916,
"learning_rate": 0.00017745030601987337,
"loss": 0.3354,
"step": 218
},
{
"epoch": 0.023450048185030516,
"grad_norm": 0.12538433074951172,
"learning_rate": 0.00017725119636778644,
"loss": 0.1759,
"step": 219
},
{
"epoch": 0.023557126030624265,
"grad_norm": 0.1331990659236908,
"learning_rate": 0.00017705132427757895,
"loss": 0.2875,
"step": 220
},
{
"epoch": 0.02366420387621801,
"grad_norm": 0.18505120277404785,
"learning_rate": 0.00017685069172190766,
"loss": 0.4,
"step": 221
},
{
"epoch": 0.023771281721811758,
"grad_norm": 0.17060880362987518,
"learning_rate": 0.00017664930068093498,
"loss": 0.4008,
"step": 222
},
{
"epoch": 0.023878359567405503,
"grad_norm": 0.151236429810524,
"learning_rate": 0.00017644715314230918,
"loss": 0.3216,
"step": 223
},
{
"epoch": 0.02398543741299925,
"grad_norm": 0.16337113082408905,
"learning_rate": 0.0001762442511011448,
"loss": 0.3807,
"step": 224
},
{
"epoch": 0.024092515258592996,
"grad_norm": 0.18361254036426544,
"learning_rate": 0.0001760405965600031,
"loss": 0.4019,
"step": 225
},
{
"epoch": 0.024199593104186744,
"grad_norm": 0.1295214742422104,
"learning_rate": 0.0001758361915288722,
"loss": 0.2674,
"step": 226
},
{
"epoch": 0.02430667094978049,
"grad_norm": 0.1772903949022293,
"learning_rate": 0.0001756310380251472,
"loss": 0.3025,
"step": 227
},
{
"epoch": 0.024413748795374238,
"grad_norm": 0.20191480219364166,
"learning_rate": 0.00017542513807361037,
"loss": 0.3619,
"step": 228
},
{
"epoch": 0.024520826640967983,
"grad_norm": 0.20930209755897522,
"learning_rate": 0.00017521849370641114,
"loss": 0.2344,
"step": 229
},
{
"epoch": 0.02462790448656173,
"grad_norm": 0.18147680163383484,
"learning_rate": 0.00017501110696304596,
"loss": 0.3712,
"step": 230
},
{
"epoch": 0.024734982332155476,
"grad_norm": 0.17919720709323883,
"learning_rate": 0.00017480297989033825,
"loss": 0.3728,
"step": 231
},
{
"epoch": 0.024842060177749224,
"grad_norm": 0.16749288141727448,
"learning_rate": 0.00017459411454241822,
"loss": 0.3647,
"step": 232
},
{
"epoch": 0.02494913802334297,
"grad_norm": 0.14664986729621887,
"learning_rate": 0.00017438451298070252,
"loss": 0.1905,
"step": 233
},
{
"epoch": 0.025056215868936717,
"grad_norm": 0.1727173924446106,
"learning_rate": 0.00017417417727387394,
"loss": 0.3138,
"step": 234
},
{
"epoch": 0.025163293714530462,
"grad_norm": 0.16467280685901642,
"learning_rate": 0.000173963109497861,
"loss": 0.2621,
"step": 235
},
{
"epoch": 0.02527037156012421,
"grad_norm": 0.21190643310546875,
"learning_rate": 0.0001737513117358174,
"loss": 0.4417,
"step": 236
},
{
"epoch": 0.025377449405717956,
"grad_norm": 0.191368967294693,
"learning_rate": 0.0001735387860781016,
"loss": 0.3023,
"step": 237
},
{
"epoch": 0.025484527251311704,
"grad_norm": 0.17699618637561798,
"learning_rate": 0.00017332553462225602,
"loss": 0.3652,
"step": 238
},
{
"epoch": 0.02559160509690545,
"grad_norm": 0.1797989159822464,
"learning_rate": 0.00017311155947298643,
"loss": 0.4054,
"step": 239
},
{
"epoch": 0.025698682942499197,
"grad_norm": 0.16934357583522797,
"learning_rate": 0.00017289686274214118,
"loss": 0.2369,
"step": 240
},
{
"epoch": 0.025805760788092942,
"grad_norm": 0.1936860829591751,
"learning_rate": 0.0001726814465486903,
"loss": 0.3687,
"step": 241
},
{
"epoch": 0.02591283863368669,
"grad_norm": 0.19604068994522095,
"learning_rate": 0.0001724653130187047,
"loss": 0.409,
"step": 242
},
{
"epoch": 0.026019916479280435,
"grad_norm": 0.16098040342330933,
"learning_rate": 0.00017224846428533499,
"loss": 0.3661,
"step": 243
},
{
"epoch": 0.026126994324874184,
"grad_norm": 0.1543160378932953,
"learning_rate": 0.0001720309024887907,
"loss": 0.286,
"step": 244
},
{
"epoch": 0.02623407217046793,
"grad_norm": 0.14484800398349762,
"learning_rate": 0.00017181262977631888,
"loss": 0.277,
"step": 245
},
{
"epoch": 0.026341150016061677,
"grad_norm": 0.18634939193725586,
"learning_rate": 0.00017159364830218312,
"loss": 0.2708,
"step": 246
},
{
"epoch": 0.026448227861655422,
"grad_norm": 0.14909416437149048,
"learning_rate": 0.00017137396022764214,
"loss": 0.2739,
"step": 247
},
{
"epoch": 0.02655530570724917,
"grad_norm": 0.17528975009918213,
"learning_rate": 0.00017115356772092857,
"loss": 0.3052,
"step": 248
},
{
"epoch": 0.026662383552842915,
"grad_norm": 0.17353974282741547,
"learning_rate": 0.0001709324729572274,
"loss": 0.3065,
"step": 249
},
{
"epoch": 0.026769461398436663,
"grad_norm": 0.15688441693782806,
"learning_rate": 0.00017071067811865476,
"loss": 0.3088,
"step": 250
},
{
"epoch": 0.026876539244030412,
"grad_norm": 0.18625454604625702,
"learning_rate": 0.00017048818539423615,
"loss": 0.4439,
"step": 251
},
{
"epoch": 0.026983617089624157,
"grad_norm": 0.17043183743953705,
"learning_rate": 0.00017026499697988493,
"loss": 0.37,
"step": 252
},
{
"epoch": 0.027090694935217905,
"grad_norm": 0.21933884918689728,
"learning_rate": 0.00017004111507838064,
"loss": 0.4339,
"step": 253
},
{
"epoch": 0.02719777278081165,
"grad_norm": 0.21618066728115082,
"learning_rate": 0.00016981654189934727,
"loss": 0.3882,
"step": 254
},
{
"epoch": 0.0273048506264054,
"grad_norm": 0.16221043467521667,
"learning_rate": 0.00016959127965923142,
"loss": 0.2924,
"step": 255
},
{
"epoch": 0.027411928471999143,
"grad_norm": 0.14445362985134125,
"learning_rate": 0.0001693653305812805,
"loss": 0.2702,
"step": 256
},
{
"epoch": 0.02751900631759289,
"grad_norm": 0.15952937304973602,
"learning_rate": 0.00016913869689552064,
"loss": 0.3569,
"step": 257
},
{
"epoch": 0.027626084163186636,
"grad_norm": 0.1786893904209137,
"learning_rate": 0.00016891138083873487,
"loss": 0.4076,
"step": 258
},
{
"epoch": 0.027733162008780385,
"grad_norm": 0.16461695730686188,
"learning_rate": 0.00016868338465444085,
"loss": 0.3617,
"step": 259
},
{
"epoch": 0.02784023985437413,
"grad_norm": 0.19281861186027527,
"learning_rate": 0.00016845471059286887,
"loss": 0.3202,
"step": 260
},
{
"epoch": 0.027947317699967878,
"grad_norm": 0.15104186534881592,
"learning_rate": 0.00016822536091093965,
"loss": 0.3157,
"step": 261
},
{
"epoch": 0.028054395545561623,
"grad_norm": 0.1911495327949524,
"learning_rate": 0.00016799533787224192,
"loss": 0.385,
"step": 262
},
{
"epoch": 0.02816147339115537,
"grad_norm": 0.17329667508602142,
"learning_rate": 0.00016776464374701025,
"loss": 0.3613,
"step": 263
},
{
"epoch": 0.028268551236749116,
"grad_norm": 0.15046873688697815,
"learning_rate": 0.00016753328081210245,
"loss": 0.2792,
"step": 264
},
{
"epoch": 0.028375629082342865,
"grad_norm": 0.1259920299053192,
"learning_rate": 0.00016730125135097735,
"loss": 0.2741,
"step": 265
},
{
"epoch": 0.02848270692793661,
"grad_norm": 0.19736653566360474,
"learning_rate": 0.000167068557653672,
"loss": 0.3415,
"step": 266
},
{
"epoch": 0.028589784773530358,
"grad_norm": 0.12615585327148438,
"learning_rate": 0.0001668352020167793,
"loss": 0.2148,
"step": 267
},
{
"epoch": 0.028696862619124103,
"grad_norm": 0.13925912976264954,
"learning_rate": 0.00016660118674342517,
"loss": 0.2772,
"step": 268
},
{
"epoch": 0.02880394046471785,
"grad_norm": 0.15701375901699066,
"learning_rate": 0.00016636651414324587,
"loss": 0.3165,
"step": 269
},
{
"epoch": 0.028911018310311596,
"grad_norm": 0.1971178501844406,
"learning_rate": 0.00016613118653236518,
"loss": 0.336,
"step": 270
},
{
"epoch": 0.029018096155905344,
"grad_norm": 0.19260185956954956,
"learning_rate": 0.0001658952062333717,
"loss": 0.3047,
"step": 271
},
{
"epoch": 0.02912517400149909,
"grad_norm": 0.17433519661426544,
"learning_rate": 0.00016565857557529566,
"loss": 0.3392,
"step": 272
},
{
"epoch": 0.029232251847092838,
"grad_norm": 0.18723376095294952,
"learning_rate": 0.00016542129689358612,
"loss": 0.3043,
"step": 273
},
{
"epoch": 0.029339329692686583,
"grad_norm": 0.12574511766433716,
"learning_rate": 0.0001651833725300879,
"loss": 0.2972,
"step": 274
},
{
"epoch": 0.02944640753828033,
"grad_norm": 0.14964695274829865,
"learning_rate": 0.00016494480483301836,
"loss": 0.3088,
"step": 275
},
{
"epoch": 0.029553485383874076,
"grad_norm": 0.17207567393779755,
"learning_rate": 0.00016470559615694446,
"loss": 0.3842,
"step": 276
},
{
"epoch": 0.029660563229467824,
"grad_norm": 0.12991392612457275,
"learning_rate": 0.00016446574886275913,
"loss": 0.2479,
"step": 277
},
{
"epoch": 0.02976764107506157,
"grad_norm": 0.14477019011974335,
"learning_rate": 0.00016422526531765846,
"loss": 0.3189,
"step": 278
},
{
"epoch": 0.029874718920655317,
"grad_norm": 0.20534910261631012,
"learning_rate": 0.00016398414789511786,
"loss": 0.3989,
"step": 279
},
{
"epoch": 0.029981796766249062,
"grad_norm": 0.14957323670387268,
"learning_rate": 0.000163742398974869,
"loss": 0.265,
"step": 280
},
{
"epoch": 0.03008887461184281,
"grad_norm": 0.17124171555042267,
"learning_rate": 0.00016350002094287609,
"loss": 0.3954,
"step": 281
},
{
"epoch": 0.030195952457436556,
"grad_norm": 0.15740445256233215,
"learning_rate": 0.00016325701619131246,
"loss": 0.3357,
"step": 282
},
{
"epoch": 0.030303030303030304,
"grad_norm": 0.1666688174009323,
"learning_rate": 0.00016301338711853693,
"loss": 0.3495,
"step": 283
},
{
"epoch": 0.03041010814862405,
"grad_norm": 0.18019483983516693,
"learning_rate": 0.00016276913612907007,
"loss": 0.4245,
"step": 284
},
{
"epoch": 0.030517185994217797,
"grad_norm": 0.21450522541999817,
"learning_rate": 0.00016252426563357055,
"loss": 0.4582,
"step": 285
},
{
"epoch": 0.030624263839811542,
"grad_norm": 0.21052783727645874,
"learning_rate": 0.00016227877804881127,
"loss": 0.4128,
"step": 286
},
{
"epoch": 0.03073134168540529,
"grad_norm": 0.1664145290851593,
"learning_rate": 0.00016203267579765563,
"loss": 0.279,
"step": 287
},
{
"epoch": 0.030838419530999035,
"grad_norm": 0.2330683022737503,
"learning_rate": 0.00016178596130903344,
"loss": 0.3742,
"step": 288
},
{
"epoch": 0.030945497376592784,
"grad_norm": 0.18258881568908691,
"learning_rate": 0.00016153863701791717,
"loss": 0.3266,
"step": 289
},
{
"epoch": 0.03105257522218653,
"grad_norm": 0.16531600058078766,
"learning_rate": 0.00016129070536529766,
"loss": 0.2748,
"step": 290
},
{
"epoch": 0.031159653067780277,
"grad_norm": 0.21928264200687408,
"learning_rate": 0.00016104216879816026,
"loss": 0.3977,
"step": 291
},
{
"epoch": 0.03126673091337402,
"grad_norm": 0.1794746071100235,
"learning_rate": 0.00016079302976946055,
"loss": 0.3939,
"step": 292
},
{
"epoch": 0.03137380875896777,
"grad_norm": 0.14978636801242828,
"learning_rate": 0.00016054329073810015,
"loss": 0.2803,
"step": 293
},
{
"epoch": 0.03148088660456152,
"grad_norm": 0.1669725626707077,
"learning_rate": 0.00016029295416890248,
"loss": 0.2965,
"step": 294
},
{
"epoch": 0.03158796445015526,
"grad_norm": 0.15341994166374207,
"learning_rate": 0.00016004202253258842,
"loss": 0.2785,
"step": 295
},
{
"epoch": 0.03169504229574901,
"grad_norm": 0.13306757807731628,
"learning_rate": 0.0001597904983057519,
"loss": 0.2762,
"step": 296
},
{
"epoch": 0.03180212014134275,
"grad_norm": 0.19124022126197815,
"learning_rate": 0.00015953838397083552,
"loss": 0.2643,
"step": 297
},
{
"epoch": 0.031909197986936505,
"grad_norm": 0.17234952747821808,
"learning_rate": 0.00015928568201610595,
"loss": 0.2988,
"step": 298
},
{
"epoch": 0.03201627583253025,
"grad_norm": 0.19267751276493073,
"learning_rate": 0.00015903239493562948,
"loss": 0.4104,
"step": 299
},
{
"epoch": 0.032123353678123995,
"grad_norm": 0.17110766470432281,
"learning_rate": 0.00015877852522924732,
"loss": 0.3486,
"step": 300
},
{
"epoch": 0.03223043152371774,
"grad_norm": 0.16783058643341064,
"learning_rate": 0.00015852407540255104,
"loss": 0.3042,
"step": 301
},
{
"epoch": 0.03233750936931149,
"grad_norm": 0.15925949811935425,
"learning_rate": 0.00015826904796685762,
"loss": 0.3339,
"step": 302
},
{
"epoch": 0.032444587214905236,
"grad_norm": 0.17646819353103638,
"learning_rate": 0.00015801344543918495,
"loss": 0.3925,
"step": 303
},
{
"epoch": 0.03255166506049898,
"grad_norm": 0.1683570295572281,
"learning_rate": 0.00015775727034222675,
"loss": 0.4217,
"step": 304
},
{
"epoch": 0.032658742906092726,
"grad_norm": 0.17200106382369995,
"learning_rate": 0.00015750052520432787,
"loss": 0.3305,
"step": 305
},
{
"epoch": 0.03276582075168648,
"grad_norm": 0.15256130695343018,
"learning_rate": 0.0001572432125594591,
"loss": 0.288,
"step": 306
},
{
"epoch": 0.03287289859728022,
"grad_norm": 0.16523954272270203,
"learning_rate": 0.00015698533494719238,
"loss": 0.3902,
"step": 307
},
{
"epoch": 0.03297997644287397,
"grad_norm": 0.12929733097553253,
"learning_rate": 0.00015672689491267567,
"loss": 0.242,
"step": 308
},
{
"epoch": 0.03308705428846771,
"grad_norm": 0.1565142422914505,
"learning_rate": 0.00015646789500660773,
"loss": 0.3979,
"step": 309
},
{
"epoch": 0.033194132134061465,
"grad_norm": 0.17319612205028534,
"learning_rate": 0.00015620833778521307,
"loss": 0.4318,
"step": 310
},
{
"epoch": 0.03330120997965521,
"grad_norm": 0.18638555705547333,
"learning_rate": 0.0001559482258102167,
"loss": 0.3411,
"step": 311
},
{
"epoch": 0.033408287825248954,
"grad_norm": 0.1663195937871933,
"learning_rate": 0.00015568756164881882,
"loss": 0.3898,
"step": 312
},
{
"epoch": 0.0335153656708427,
"grad_norm": 0.12741482257843018,
"learning_rate": 0.00015542634787366942,
"loss": 0.2005,
"step": 313
},
{
"epoch": 0.03362244351643645,
"grad_norm": 0.14058327674865723,
"learning_rate": 0.00015516458706284303,
"loss": 0.2665,
"step": 314
},
{
"epoch": 0.033729521362030196,
"grad_norm": 0.14359046518802643,
"learning_rate": 0.0001549022817998132,
"loss": 0.3479,
"step": 315
},
{
"epoch": 0.03383659920762394,
"grad_norm": 0.17742228507995605,
"learning_rate": 0.00015463943467342693,
"loss": 0.4099,
"step": 316
},
{
"epoch": 0.03394367705321769,
"grad_norm": 0.12765684723854065,
"learning_rate": 0.00015437604827787927,
"loss": 0.2812,
"step": 317
},
{
"epoch": 0.03405075489881144,
"grad_norm": 0.1804419457912445,
"learning_rate": 0.00015411212521268758,
"loss": 0.2586,
"step": 318
},
{
"epoch": 0.03415783274440518,
"grad_norm": 0.21663552522659302,
"learning_rate": 0.00015384766808266602,
"loss": 0.3947,
"step": 319
},
{
"epoch": 0.03426491058999893,
"grad_norm": 0.15343870222568512,
"learning_rate": 0.00015358267949789966,
"loss": 0.3475,
"step": 320
},
{
"epoch": 0.03437198843559268,
"grad_norm": 0.1694641411304474,
"learning_rate": 0.00015331716207371888,
"loss": 0.36,
"step": 321
},
{
"epoch": 0.034479066281186424,
"grad_norm": 0.24741993844509125,
"learning_rate": 0.0001530511184306734,
"loss": 0.3511,
"step": 322
},
{
"epoch": 0.03458614412678017,
"grad_norm": 0.17976830899715424,
"learning_rate": 0.00015278455119450664,
"loss": 0.3648,
"step": 323
},
{
"epoch": 0.034693221972373914,
"grad_norm": 0.1624085009098053,
"learning_rate": 0.0001525174629961296,
"loss": 0.399,
"step": 324
},
{
"epoch": 0.034800299817967666,
"grad_norm": 0.14293444156646729,
"learning_rate": 0.0001522498564715949,
"loss": 0.3321,
"step": 325
},
{
"epoch": 0.03490737766356141,
"grad_norm": 0.1629532128572464,
"learning_rate": 0.00015198173426207094,
"loss": 0.3283,
"step": 326
},
{
"epoch": 0.035014455509155155,
"grad_norm": 0.1322093904018402,
"learning_rate": 0.00015171309901381572,
"loss": 0.2678,
"step": 327
},
{
"epoch": 0.0351215333547489,
"grad_norm": 0.17138321697711945,
"learning_rate": 0.00015144395337815064,
"loss": 0.2985,
"step": 328
},
{
"epoch": 0.03522861120034265,
"grad_norm": 0.17492185533046722,
"learning_rate": 0.00015117430001143452,
"loss": 0.3936,
"step": 329
},
{
"epoch": 0.0353356890459364,
"grad_norm": 0.15431442856788635,
"learning_rate": 0.00015090414157503714,
"loss": 0.2784,
"step": 330
},
{
"epoch": 0.03544276689153014,
"grad_norm": 0.14638526737689972,
"learning_rate": 0.00015063348073531324,
"loss": 0.2342,
"step": 331
},
{
"epoch": 0.03554984473712389,
"grad_norm": 0.14522488415241241,
"learning_rate": 0.0001503623201635761,
"loss": 0.2721,
"step": 332
},
{
"epoch": 0.03565692258271764,
"grad_norm": 0.1473628282546997,
"learning_rate": 0.000150090662536071,
"loss": 0.2752,
"step": 333
},
{
"epoch": 0.035764000428311384,
"grad_norm": 0.18698623776435852,
"learning_rate": 0.0001498185105339491,
"loss": 0.3196,
"step": 334
},
{
"epoch": 0.03587107827390513,
"grad_norm": 0.16111692786216736,
"learning_rate": 0.00014954586684324078,
"loss": 0.3039,
"step": 335
},
{
"epoch": 0.03597815611949887,
"grad_norm": 0.16252991557121277,
"learning_rate": 0.00014927273415482915,
"loss": 0.2696,
"step": 336
},
{
"epoch": 0.036085233965092625,
"grad_norm": 0.16255027055740356,
"learning_rate": 0.00014899911516442365,
"loss": 0.2249,
"step": 337
},
{
"epoch": 0.03619231181068637,
"grad_norm": 0.14544181525707245,
"learning_rate": 0.00014872501257253323,
"loss": 0.2746,
"step": 338
},
{
"epoch": 0.036299389656280115,
"grad_norm": 0.13909004628658295,
"learning_rate": 0.0001484504290844398,
"loss": 0.2982,
"step": 339
},
{
"epoch": 0.03640646750187386,
"grad_norm": 0.20055408775806427,
"learning_rate": 0.00014817536741017152,
"loss": 0.357,
"step": 340
},
{
"epoch": 0.03651354534746761,
"grad_norm": 0.1298537254333496,
"learning_rate": 0.00014789983026447612,
"loss": 0.2204,
"step": 341
},
{
"epoch": 0.03662062319306136,
"grad_norm": 0.19047404825687408,
"learning_rate": 0.0001476238203667939,
"loss": 0.2952,
"step": 342
},
{
"epoch": 0.0367277010386551,
"grad_norm": 0.18302544951438904,
"learning_rate": 0.0001473473404412312,
"loss": 0.2872,
"step": 343
},
{
"epoch": 0.036834778884248846,
"grad_norm": 0.1702168881893158,
"learning_rate": 0.0001470703932165333,
"loss": 0.332,
"step": 344
},
{
"epoch": 0.0369418567298426,
"grad_norm": 0.20581865310668945,
"learning_rate": 0.00014679298142605734,
"loss": 0.4375,
"step": 345
},
{
"epoch": 0.03704893457543634,
"grad_norm": 0.17323224246501923,
"learning_rate": 0.00014651510780774583,
"loss": 0.3551,
"step": 346
},
{
"epoch": 0.03715601242103009,
"grad_norm": 0.149403378367424,
"learning_rate": 0.00014623677510409918,
"loss": 0.2716,
"step": 347
},
{
"epoch": 0.03726309026662383,
"grad_norm": 0.15081925690174103,
"learning_rate": 0.00014595798606214882,
"loss": 0.311,
"step": 348
},
{
"epoch": 0.037370168112217585,
"grad_norm": 0.18388235569000244,
"learning_rate": 0.00014567874343342997,
"loss": 0.36,
"step": 349
},
{
"epoch": 0.03747724595781133,
"grad_norm": 0.17369544506072998,
"learning_rate": 0.00014539904997395468,
"loss": 0.3535,
"step": 350
},
{
"epoch": 0.037584323803405074,
"grad_norm": 0.17372484505176544,
"learning_rate": 0.00014511890844418453,
"loss": 0.2509,
"step": 351
},
{
"epoch": 0.03769140164899882,
"grad_norm": 0.1707952469587326,
"learning_rate": 0.00014483832160900326,
"loss": 0.4087,
"step": 352
},
{
"epoch": 0.03779847949459257,
"grad_norm": 0.18640413880348206,
"learning_rate": 0.00014455729223768966,
"loss": 0.3737,
"step": 353
},
{
"epoch": 0.037905557340186316,
"grad_norm": 0.1838582456111908,
"learning_rate": 0.0001442758231038902,
"loss": 0.4719,
"step": 354
},
{
"epoch": 0.03801263518578006,
"grad_norm": 0.16711768507957458,
"learning_rate": 0.00014399391698559152,
"loss": 0.303,
"step": 355
},
{
"epoch": 0.038119713031373806,
"grad_norm": 0.209504172205925,
"learning_rate": 0.0001437115766650933,
"loss": 0.3234,
"step": 356
},
{
"epoch": 0.03822679087696756,
"grad_norm": 0.19139063358306885,
"learning_rate": 0.00014342880492898048,
"loss": 0.2605,
"step": 357
},
{
"epoch": 0.0383338687225613,
"grad_norm": 0.14143365621566772,
"learning_rate": 0.0001431456045680959,
"loss": 0.3509,
"step": 358
},
{
"epoch": 0.03844094656815505,
"grad_norm": 0.14169888198375702,
"learning_rate": 0.00014286197837751286,
"loss": 0.2654,
"step": 359
},
{
"epoch": 0.03854802441374879,
"grad_norm": 0.14783132076263428,
"learning_rate": 0.00014257792915650728,
"loss": 0.2976,
"step": 360
},
{
"epoch": 0.038655102259342544,
"grad_norm": 0.14012813568115234,
"learning_rate": 0.00014229345970853032,
"loss": 0.2406,
"step": 361
},
{
"epoch": 0.03876218010493629,
"grad_norm": 0.20082809031009674,
"learning_rate": 0.00014200857284118066,
"loss": 0.3533,
"step": 362
},
{
"epoch": 0.038869257950530034,
"grad_norm": 0.1798025518655777,
"learning_rate": 0.00014172327136617656,
"loss": 0.3245,
"step": 363
},
{
"epoch": 0.03897633579612378,
"grad_norm": 0.16510576009750366,
"learning_rate": 0.00014143755809932845,
"loss": 0.2857,
"step": 364
},
{
"epoch": 0.03908341364171753,
"grad_norm": 0.16456517577171326,
"learning_rate": 0.00014115143586051088,
"loss": 0.3673,
"step": 365
},
{
"epoch": 0.039190491487311276,
"grad_norm": 0.1842593401670456,
"learning_rate": 0.00014086490747363493,
"loss": 0.3667,
"step": 366
},
{
"epoch": 0.03929756933290502,
"grad_norm": 0.18094277381896973,
"learning_rate": 0.00014057797576662,
"loss": 0.394,
"step": 367
},
{
"epoch": 0.039404647178498765,
"grad_norm": 0.14859478175640106,
"learning_rate": 0.00014029064357136628,
"loss": 0.2738,
"step": 368
},
{
"epoch": 0.03951172502409252,
"grad_norm": 0.16508136689662933,
"learning_rate": 0.00014000291372372647,
"loss": 0.3326,
"step": 369
},
{
"epoch": 0.03961880286968626,
"grad_norm": 0.16775007545948029,
"learning_rate": 0.00013971478906347806,
"loss": 0.4139,
"step": 370
},
{
"epoch": 0.03972588071528001,
"grad_norm": 0.1426904797554016,
"learning_rate": 0.00013942627243429512,
"loss": 0.2177,
"step": 371
},
{
"epoch": 0.03983295856087375,
"grad_norm": 0.20198000967502594,
"learning_rate": 0.00013913736668372026,
"loss": 0.453,
"step": 372
},
{
"epoch": 0.039940036406467504,
"grad_norm": 0.1864636093378067,
"learning_rate": 0.00013884807466313663,
"loss": 0.3124,
"step": 373
},
{
"epoch": 0.04004711425206125,
"grad_norm": 0.13631337881088257,
"learning_rate": 0.00013855839922773968,
"loss": 0.262,
"step": 374
},
{
"epoch": 0.040154192097654993,
"grad_norm": 0.17490577697753906,
"learning_rate": 0.000138268343236509,
"loss": 0.3139,
"step": 375
},
{
"epoch": 0.04026126994324874,
"grad_norm": 0.1761641502380371,
"learning_rate": 0.00013797790955218014,
"loss": 0.4065,
"step": 376
},
{
"epoch": 0.04036834778884249,
"grad_norm": 0.156477689743042,
"learning_rate": 0.00013768710104121627,
"loss": 0.2875,
"step": 377
},
{
"epoch": 0.040475425634436235,
"grad_norm": 0.14239205420017242,
"learning_rate": 0.00013739592057378003,
"loss": 0.3407,
"step": 378
},
{
"epoch": 0.04058250348002998,
"grad_norm": 0.18152830004692078,
"learning_rate": 0.0001371043710237051,
"loss": 0.3821,
"step": 379
},
{
"epoch": 0.04068958132562373,
"grad_norm": 0.14406077563762665,
"learning_rate": 0.00013681245526846783,
"loss": 0.286,
"step": 380
},
{
"epoch": 0.04079665917121748,
"grad_norm": 0.1428261250257492,
"learning_rate": 0.0001365201761891588,
"loss": 0.2463,
"step": 381
},
{
"epoch": 0.04090373701681122,
"grad_norm": 0.16200338304042816,
"learning_rate": 0.00013622753667045457,
"loss": 0.3641,
"step": 382
},
{
"epoch": 0.041010814862404966,
"grad_norm": 0.19177362322807312,
"learning_rate": 0.00013593453960058908,
"loss": 0.3988,
"step": 383
},
{
"epoch": 0.04111789270799872,
"grad_norm": 0.13588127493858337,
"learning_rate": 0.00013564118787132506,
"loss": 0.3097,
"step": 384
},
{
"epoch": 0.04122497055359246,
"grad_norm": 0.14050887525081635,
"learning_rate": 0.00013534748437792573,
"loss": 0.2883,
"step": 385
},
{
"epoch": 0.04133204839918621,
"grad_norm": 0.1991889923810959,
"learning_rate": 0.0001350534320191259,
"loss": 0.3118,
"step": 386
},
{
"epoch": 0.04143912624477995,
"grad_norm": 0.1473557949066162,
"learning_rate": 0.0001347590336971037,
"loss": 0.363,
"step": 387
},
{
"epoch": 0.041546204090373705,
"grad_norm": 0.20648524165153503,
"learning_rate": 0.0001344642923174517,
"loss": 0.3746,
"step": 388
},
{
"epoch": 0.04165328193596745,
"grad_norm": 0.12530086934566498,
"learning_rate": 0.00013416921078914835,
"loss": 0.2218,
"step": 389
},
{
"epoch": 0.041760359781561195,
"grad_norm": 0.1354837268590927,
"learning_rate": 0.00013387379202452917,
"loss": 0.2357,
"step": 390
},
{
"epoch": 0.04186743762715494,
"grad_norm": 0.13786575198173523,
"learning_rate": 0.00013357803893925807,
"loss": 0.2601,
"step": 391
},
{
"epoch": 0.04197451547274869,
"grad_norm": 0.14999490976333618,
"learning_rate": 0.00013328195445229868,
"loss": 0.2587,
"step": 392
},
{
"epoch": 0.042081593318342436,
"grad_norm": 0.20544356107711792,
"learning_rate": 0.00013298554148588528,
"loss": 0.3826,
"step": 393
},
{
"epoch": 0.04218867116393618,
"grad_norm": 0.18927697837352753,
"learning_rate": 0.00013268880296549425,
"loss": 0.3572,
"step": 394
},
{
"epoch": 0.042295749009529926,
"grad_norm": 0.18653343617916107,
"learning_rate": 0.00013239174181981495,
"loss": 0.3651,
"step": 395
},
{
"epoch": 0.04240282685512368,
"grad_norm": 0.2660796642303467,
"learning_rate": 0.00013209436098072095,
"loss": 0.4001,
"step": 396
},
{
"epoch": 0.04250990470071742,
"grad_norm": 0.1991802304983139,
"learning_rate": 0.00013179666338324108,
"loss": 0.399,
"step": 397
},
{
"epoch": 0.04261698254631117,
"grad_norm": 0.20211124420166016,
"learning_rate": 0.0001314986519655305,
"loss": 0.452,
"step": 398
},
{
"epoch": 0.04272406039190491,
"grad_norm": 0.19349808990955353,
"learning_rate": 0.0001312003296688415,
"loss": 0.5259,
"step": 399
},
{
"epoch": 0.042831138237498664,
"grad_norm": 0.1513546109199524,
"learning_rate": 0.00013090169943749476,
"loss": 0.2608,
"step": 400
},
{
"epoch": 0.04293821608309241,
"grad_norm": 0.15879368782043457,
"learning_rate": 0.0001306027642188501,
"loss": 0.3402,
"step": 401
},
{
"epoch": 0.043045293928686154,
"grad_norm": 0.15345223248004913,
"learning_rate": 0.00013030352696327742,
"loss": 0.311,
"step": 402
},
{
"epoch": 0.0431523717742799,
"grad_norm": 0.17926479876041412,
"learning_rate": 0.00013000399062412763,
"loss": 0.3439,
"step": 403
},
{
"epoch": 0.04325944961987365,
"grad_norm": 0.1882927417755127,
"learning_rate": 0.0001297041581577035,
"loss": 0.3147,
"step": 404
},
{
"epoch": 0.043366527465467396,
"grad_norm": 0.12865504622459412,
"learning_rate": 0.0001294040325232304,
"loss": 0.3,
"step": 405
},
{
"epoch": 0.04347360531106114,
"grad_norm": 0.15574470162391663,
"learning_rate": 0.00012910361668282719,
"loss": 0.3666,
"step": 406
},
{
"epoch": 0.043580683156654886,
"grad_norm": 0.19106611609458923,
"learning_rate": 0.00012880291360147693,
"loss": 0.3568,
"step": 407
},
{
"epoch": 0.04368776100224864,
"grad_norm": 0.18762201070785522,
"learning_rate": 0.0001285019262469976,
"loss": 0.3123,
"step": 408
},
{
"epoch": 0.04379483884784238,
"grad_norm": 0.14002811908721924,
"learning_rate": 0.00012820065759001293,
"loss": 0.2737,
"step": 409
},
{
"epoch": 0.04390191669343613,
"grad_norm": 0.19170932471752167,
"learning_rate": 0.00012789911060392294,
"loss": 0.4061,
"step": 410
},
{
"epoch": 0.04400899453902987,
"grad_norm": 0.17306004464626312,
"learning_rate": 0.0001275972882648746,
"loss": 0.3492,
"step": 411
},
{
"epoch": 0.044116072384623624,
"grad_norm": 0.1948164999485016,
"learning_rate": 0.00012729519355173254,
"loss": 0.4394,
"step": 412
},
{
"epoch": 0.04422315023021737,
"grad_norm": 0.15455390512943268,
"learning_rate": 0.00012699282944604967,
"loss": 0.3778,
"step": 413
},
{
"epoch": 0.044330228075811114,
"grad_norm": 0.1604427546262741,
"learning_rate": 0.00012669019893203759,
"loss": 0.3684,
"step": 414
},
{
"epoch": 0.04443730592140486,
"grad_norm": 0.16993975639343262,
"learning_rate": 0.0001263873049965373,
"loss": 0.2551,
"step": 415
},
{
"epoch": 0.04454438376699861,
"grad_norm": 0.15334467589855194,
"learning_rate": 0.00012608415062898972,
"loss": 0.3069,
"step": 416
},
{
"epoch": 0.044651461612592355,
"grad_norm": 0.1586979478597641,
"learning_rate": 0.000125780738821406,
"loss": 0.3423,
"step": 417
},
{
"epoch": 0.0447585394581861,
"grad_norm": 0.1945161521434784,
"learning_rate": 0.00012547707256833823,
"loss": 0.3957,
"step": 418
},
{
"epoch": 0.044865617303779845,
"grad_norm": 0.18735453486442566,
"learning_rate": 0.00012517315486684972,
"loss": 0.3792,
"step": 419
},
{
"epoch": 0.0449726951493736,
"grad_norm": 0.16127122938632965,
"learning_rate": 0.0001248689887164855,
"loss": 0.3564,
"step": 420
},
{
"epoch": 0.04507977299496734,
"grad_norm": 0.1495504528284073,
"learning_rate": 0.00012456457711924266,
"loss": 0.2447,
"step": 421
},
{
"epoch": 0.04518685084056109,
"grad_norm": 0.12814819812774658,
"learning_rate": 0.00012425992307954075,
"loss": 0.2419,
"step": 422
},
{
"epoch": 0.04529392868615483,
"grad_norm": 0.13747601211071014,
"learning_rate": 0.0001239550296041922,
"loss": 0.2881,
"step": 423
},
{
"epoch": 0.04540100653174858,
"grad_norm": 0.14220485091209412,
"learning_rate": 0.00012364989970237248,
"loss": 0.2969,
"step": 424
},
{
"epoch": 0.04550808437734233,
"grad_norm": 0.15737488865852356,
"learning_rate": 0.00012334453638559057,
"loss": 0.3217,
"step": 425
},
{
"epoch": 0.04561516222293607,
"grad_norm": 0.13472171127796173,
"learning_rate": 0.00012303894266765908,
"loss": 0.2394,
"step": 426
},
{
"epoch": 0.04572224006852982,
"grad_norm": 0.16132071614265442,
"learning_rate": 0.00012273312156466464,
"loss": 0.2786,
"step": 427
},
{
"epoch": 0.04582931791412357,
"grad_norm": 0.17469573020935059,
"learning_rate": 0.00012242707609493814,
"loss": 0.2979,
"step": 428
},
{
"epoch": 0.045936395759717315,
"grad_norm": 0.1818244755268097,
"learning_rate": 0.00012212080927902474,
"loss": 0.3639,
"step": 429
},
{
"epoch": 0.04604347360531106,
"grad_norm": 0.1542896330356598,
"learning_rate": 0.00012181432413965428,
"loss": 0.2892,
"step": 430
},
{
"epoch": 0.046150551450904805,
"grad_norm": 0.17613442242145538,
"learning_rate": 0.00012150762370171136,
"loss": 0.424,
"step": 431
},
{
"epoch": 0.046257629296498556,
"grad_norm": 0.17226651310920715,
"learning_rate": 0.00012120071099220549,
"loss": 0.3764,
"step": 432
},
{
"epoch": 0.0463647071420923,
"grad_norm": 0.16670876741409302,
"learning_rate": 0.00012089358904024117,
"loss": 0.3949,
"step": 433
},
{
"epoch": 0.046471784987686046,
"grad_norm": 0.137644961476326,
"learning_rate": 0.00012058626087698814,
"loss": 0.203,
"step": 434
},
{
"epoch": 0.04657886283327979,
"grad_norm": 0.18664462864398956,
"learning_rate": 0.00012027872953565125,
"loss": 0.3211,
"step": 435
},
{
"epoch": 0.04668594067887354,
"grad_norm": 0.2129397839307785,
"learning_rate": 0.00011997099805144069,
"loss": 0.4424,
"step": 436
},
{
"epoch": 0.04679301852446729,
"grad_norm": 0.1846884787082672,
"learning_rate": 0.000119663069461542,
"loss": 0.2833,
"step": 437
},
{
"epoch": 0.04690009637006103,
"grad_norm": 0.1526833325624466,
"learning_rate": 0.00011935494680508606,
"loss": 0.3097,
"step": 438
},
{
"epoch": 0.047007174215654784,
"grad_norm": 0.20358015596866608,
"learning_rate": 0.00011904663312311901,
"loss": 0.3995,
"step": 439
},
{
"epoch": 0.04711425206124853,
"grad_norm": 0.21734708547592163,
"learning_rate": 0.00011873813145857249,
"loss": 0.3412,
"step": 440
},
{
"epoch": 0.047221329906842274,
"grad_norm": 0.18342752754688263,
"learning_rate": 0.00011842944485623335,
"loss": 0.336,
"step": 441
},
{
"epoch": 0.04732840775243602,
"grad_norm": 0.16440509259700775,
"learning_rate": 0.00011812057636271374,
"loss": 0.2243,
"step": 442
},
{
"epoch": 0.04743548559802977,
"grad_norm": 0.16069766879081726,
"learning_rate": 0.000117811529026421,
"loss": 0.3723,
"step": 443
},
{
"epoch": 0.047542563443623516,
"grad_norm": 0.16657786071300507,
"learning_rate": 0.00011750230589752762,
"loss": 0.3583,
"step": 444
},
{
"epoch": 0.04764964128921726,
"grad_norm": 0.18991921842098236,
"learning_rate": 0.00011719291002794096,
"loss": 0.3201,
"step": 445
},
{
"epoch": 0.047756719134811006,
"grad_norm": 0.14543665945529938,
"learning_rate": 0.00011688334447127338,
"loss": 0.3762,
"step": 446
},
{
"epoch": 0.04786379698040476,
"grad_norm": 0.1752735674381256,
"learning_rate": 0.00011657361228281199,
"loss": 0.3603,
"step": 447
},
{
"epoch": 0.0479708748259985,
"grad_norm": 0.15264953672885895,
"learning_rate": 0.00011626371651948838,
"loss": 0.2599,
"step": 448
},
{
"epoch": 0.04807795267159225,
"grad_norm": 0.10569647699594498,
"learning_rate": 0.00011595366023984864,
"loss": 0.1823,
"step": 449
},
{
"epoch": 0.04818503051718599,
"grad_norm": 0.19080232083797455,
"learning_rate": 0.0001156434465040231,
"loss": 0.3381,
"step": 450
},
{
"epoch": 0.048292108362779744,
"grad_norm": 0.1927441954612732,
"learning_rate": 0.00011533307837369607,
"loss": 0.3542,
"step": 451
},
{
"epoch": 0.04839918620837349,
"grad_norm": 0.18672965466976166,
"learning_rate": 0.00011502255891207572,
"loss": 0.3396,
"step": 452
},
{
"epoch": 0.048506264053967234,
"grad_norm": 0.20121031999588013,
"learning_rate": 0.00011471189118386375,
"loss": 0.3582,
"step": 453
},
{
"epoch": 0.04861334189956098,
"grad_norm": 0.14637677371501923,
"learning_rate": 0.00011440107825522521,
"loss": 0.2784,
"step": 454
},
{
"epoch": 0.04872041974515473,
"grad_norm": 0.16313035786151886,
"learning_rate": 0.00011409012319375827,
"loss": 0.4364,
"step": 455
},
{
"epoch": 0.048827497590748475,
"grad_norm": 0.17453531920909882,
"learning_rate": 0.0001137790290684638,
"loss": 0.4113,
"step": 456
},
{
"epoch": 0.04893457543634222,
"grad_norm": 0.19787636399269104,
"learning_rate": 0.00011346779894971527,
"loss": 0.4537,
"step": 457
},
{
"epoch": 0.049041653281935965,
"grad_norm": 0.16840681433677673,
"learning_rate": 0.00011315643590922827,
"loss": 0.3678,
"step": 458
},
{
"epoch": 0.04914873112752972,
"grad_norm": 0.1894865781068802,
"learning_rate": 0.0001128449430200303,
"loss": 0.4161,
"step": 459
},
{
"epoch": 0.04925580897312346,
"grad_norm": 0.180477574467659,
"learning_rate": 0.00011253332335643043,
"loss": 0.3242,
"step": 460
},
{
"epoch": 0.04936288681871721,
"grad_norm": 0.17623566091060638,
"learning_rate": 0.00011222157999398895,
"loss": 0.2212,
"step": 461
},
{
"epoch": 0.04946996466431095,
"grad_norm": 0.14930705726146698,
"learning_rate": 0.00011190971600948699,
"loss": 0.3156,
"step": 462
},
{
"epoch": 0.049577042509904704,
"grad_norm": 0.19468611478805542,
"learning_rate": 0.00011159773448089614,
"loss": 0.4011,
"step": 463
},
{
"epoch": 0.04968412035549845,
"grad_norm": 0.17452605068683624,
"learning_rate": 0.00011128563848734816,
"loss": 0.3711,
"step": 464
},
{
"epoch": 0.04979119820109219,
"grad_norm": 0.15289977192878723,
"learning_rate": 0.00011097343110910452,
"loss": 0.3943,
"step": 465
},
{
"epoch": 0.04989827604668594,
"grad_norm": 0.19616980850696564,
"learning_rate": 0.000110661115427526,
"loss": 0.3831,
"step": 466
},
{
"epoch": 0.05000535389227969,
"grad_norm": 0.15868891775608063,
"learning_rate": 0.00011034869452504226,
"loss": 0.3308,
"step": 467
},
{
"epoch": 0.050112431737873435,
"grad_norm": 0.14652469754219055,
"learning_rate": 0.00011003617148512149,
"loss": 0.2859,
"step": 468
},
{
"epoch": 0.05021950958346718,
"grad_norm": 0.15360702574253082,
"learning_rate": 0.00010972354939223996,
"loss": 0.2889,
"step": 469
},
{
"epoch": 0.050326587429060925,
"grad_norm": 0.19247639179229736,
"learning_rate": 0.00010941083133185146,
"loss": 0.3052,
"step": 470
},
{
"epoch": 0.050433665274654677,
"grad_norm": 0.17776168882846832,
"learning_rate": 0.00010909802039035701,
"loss": 0.393,
"step": 471
},
{
"epoch": 0.05054074312024842,
"grad_norm": 0.15174326300621033,
"learning_rate": 0.00010878511965507434,
"loss": 0.2693,
"step": 472
},
{
"epoch": 0.050647820965842166,
"grad_norm": 0.17235317826271057,
"learning_rate": 0.00010847213221420736,
"loss": 0.2949,
"step": 473
},
{
"epoch": 0.05075489881143591,
"grad_norm": 0.16213464736938477,
"learning_rate": 0.00010815906115681578,
"loss": 0.3521,
"step": 474
},
{
"epoch": 0.05086197665702966,
"grad_norm": 0.18122194707393646,
"learning_rate": 0.0001078459095727845,
"loss": 0.2741,
"step": 475
},
{
"epoch": 0.05096905450262341,
"grad_norm": 0.2096056491136551,
"learning_rate": 0.00010753268055279329,
"loss": 0.3368,
"step": 476
},
{
"epoch": 0.05107613234821715,
"grad_norm": 0.148423969745636,
"learning_rate": 0.0001072193771882861,
"loss": 0.3058,
"step": 477
},
{
"epoch": 0.0511832101938109,
"grad_norm": 0.1586843580007553,
"learning_rate": 0.00010690600257144061,
"loss": 0.3129,
"step": 478
},
{
"epoch": 0.05129028803940465,
"grad_norm": 0.15282008051872253,
"learning_rate": 0.0001065925597951378,
"loss": 0.2693,
"step": 479
},
{
"epoch": 0.051397365884998394,
"grad_norm": 0.16839781403541565,
"learning_rate": 0.00010627905195293135,
"loss": 0.2877,
"step": 480
},
{
"epoch": 0.05150444373059214,
"grad_norm": 0.16241709887981415,
"learning_rate": 0.00010596548213901708,
"loss": 0.2732,
"step": 481
},
{
"epoch": 0.051611521576185884,
"grad_norm": 0.17643004655838013,
"learning_rate": 0.00010565185344820247,
"loss": 0.362,
"step": 482
},
{
"epoch": 0.051718599421779636,
"grad_norm": 0.17319577932357788,
"learning_rate": 0.00010533816897587606,
"loss": 0.3744,
"step": 483
},
{
"epoch": 0.05182567726737338,
"grad_norm": 0.1710813194513321,
"learning_rate": 0.00010502443181797697,
"loss": 0.3154,
"step": 484
},
{
"epoch": 0.051932755112967126,
"grad_norm": 0.17228226363658905,
"learning_rate": 0.00010471064507096426,
"loss": 0.3267,
"step": 485
},
{
"epoch": 0.05203983295856087,
"grad_norm": 0.16815848648548126,
"learning_rate": 0.0001043968118317865,
"loss": 0.4157,
"step": 486
},
{
"epoch": 0.05214691080415462,
"grad_norm": 0.14662803709506989,
"learning_rate": 0.00010408293519785101,
"loss": 0.2119,
"step": 487
},
{
"epoch": 0.05225398864974837,
"grad_norm": 0.14955608546733856,
"learning_rate": 0.00010376901826699348,
"loss": 0.3402,
"step": 488
},
{
"epoch": 0.05236106649534211,
"grad_norm": 0.20436280965805054,
"learning_rate": 0.00010345506413744726,
"loss": 0.2788,
"step": 489
},
{
"epoch": 0.05246814434093586,
"grad_norm": 0.1706622689962387,
"learning_rate": 0.00010314107590781284,
"loss": 0.3243,
"step": 490
},
{
"epoch": 0.05257522218652961,
"grad_norm": 0.17119503021240234,
"learning_rate": 0.00010282705667702734,
"loss": 0.2811,
"step": 491
},
{
"epoch": 0.052682300032123354,
"grad_norm": 0.145518958568573,
"learning_rate": 0.00010251300954433376,
"loss": 0.3067,
"step": 492
},
{
"epoch": 0.0527893778777171,
"grad_norm": 0.1998199224472046,
"learning_rate": 0.00010219893760925052,
"loss": 0.4282,
"step": 493
},
{
"epoch": 0.052896455723310844,
"grad_norm": 0.15826267004013062,
"learning_rate": 0.00010188484397154084,
"loss": 0.3698,
"step": 494
},
{
"epoch": 0.053003533568904596,
"grad_norm": 0.15670500695705414,
"learning_rate": 0.00010157073173118208,
"loss": 0.3849,
"step": 495
},
{
"epoch": 0.05311061141449834,
"grad_norm": 0.214689239859581,
"learning_rate": 0.00010125660398833528,
"loss": 0.4448,
"step": 496
},
{
"epoch": 0.053217689260092085,
"grad_norm": 0.17683358490467072,
"learning_rate": 0.00010094246384331442,
"loss": 0.307,
"step": 497
},
{
"epoch": 0.05332476710568583,
"grad_norm": 0.236779123544693,
"learning_rate": 0.00010062831439655591,
"loss": 0.4571,
"step": 498
},
{
"epoch": 0.05343184495127958,
"grad_norm": 0.19618524610996246,
"learning_rate": 0.00010031415874858797,
"loss": 0.5339,
"step": 499
},
{
"epoch": 0.05353892279687333,
"grad_norm": 0.14964862167835236,
"learning_rate": 0.0001,
"loss": 0.246,
"step": 500
},
{
"epoch": 0.05364600064246707,
"grad_norm": 0.15746654570102692,
"learning_rate": 9.968584125141204e-05,
"loss": 0.3532,
"step": 501
},
{
"epoch": 0.053753078488060824,
"grad_norm": 0.21315313875675201,
"learning_rate": 9.937168560344412e-05,
"loss": 0.4048,
"step": 502
},
{
"epoch": 0.05386015633365457,
"grad_norm": 0.17655318975448608,
"learning_rate": 9.90575361566856e-05,
"loss": 0.319,
"step": 503
},
{
"epoch": 0.05396723417924831,
"grad_norm": 0.15725268423557281,
"learning_rate": 9.874339601166473e-05,
"loss": 0.2798,
"step": 504
},
{
"epoch": 0.05407431202484206,
"grad_norm": 0.15933293104171753,
"learning_rate": 9.842926826881796e-05,
"loss": 0.3104,
"step": 505
},
{
"epoch": 0.05418138987043581,
"grad_norm": 0.11724121868610382,
"learning_rate": 9.81151560284592e-05,
"loss": 0.2285,
"step": 506
},
{
"epoch": 0.054288467716029555,
"grad_norm": 0.16910420358181,
"learning_rate": 9.78010623907495e-05,
"loss": 0.3072,
"step": 507
},
{
"epoch": 0.0543955455616233,
"grad_norm": 0.15870051085948944,
"learning_rate": 9.748699045566626e-05,
"loss": 0.3475,
"step": 508
},
{
"epoch": 0.054502623407217045,
"grad_norm": 0.15897218883037567,
"learning_rate": 9.717294332297268e-05,
"loss": 0.2984,
"step": 509
},
{
"epoch": 0.0546097012528108,
"grad_norm": 0.15372000634670258,
"learning_rate": 9.685892409218717e-05,
"loss": 0.3385,
"step": 510
},
{
"epoch": 0.05471677909840454,
"grad_norm": 0.18505370616912842,
"learning_rate": 9.654493586255278e-05,
"loss": 0.3844,
"step": 511
},
{
"epoch": 0.054823856943998286,
"grad_norm": 0.1893370896577835,
"learning_rate": 9.623098173300654e-05,
"loss": 0.4111,
"step": 512
},
{
"epoch": 0.05493093478959203,
"grad_norm": 0.1814618706703186,
"learning_rate": 9.591706480214901e-05,
"loss": 0.2973,
"step": 513
},
{
"epoch": 0.05503801263518578,
"grad_norm": 0.17390666902065277,
"learning_rate": 9.560318816821353e-05,
"loss": 0.3329,
"step": 514
},
{
"epoch": 0.05514509048077953,
"grad_norm": 0.14959359169006348,
"learning_rate": 9.528935492903575e-05,
"loss": 0.2716,
"step": 515
},
{
"epoch": 0.05525216832637327,
"grad_norm": 0.1462404578924179,
"learning_rate": 9.497556818202306e-05,
"loss": 0.3095,
"step": 516
},
{
"epoch": 0.05535924617196702,
"grad_norm": 0.1766779124736786,
"learning_rate": 9.466183102412395e-05,
"loss": 0.3714,
"step": 517
},
{
"epoch": 0.05546632401756077,
"grad_norm": 0.1642884612083435,
"learning_rate": 9.434814655179755e-05,
"loss": 0.2852,
"step": 518
},
{
"epoch": 0.055573401863154515,
"grad_norm": 0.16393111646175385,
"learning_rate": 9.403451786098294e-05,
"loss": 0.3584,
"step": 519
},
{
"epoch": 0.05568047970874826,
"grad_norm": 0.2010338306427002,
"learning_rate": 9.372094804706867e-05,
"loss": 0.515,
"step": 520
},
{
"epoch": 0.055787557554342004,
"grad_norm": 0.148441344499588,
"learning_rate": 9.340744020486222e-05,
"loss": 0.3691,
"step": 521
},
{
"epoch": 0.055894635399935756,
"grad_norm": 0.18052120506763458,
"learning_rate": 9.309399742855942e-05,
"loss": 0.3099,
"step": 522
},
{
"epoch": 0.0560017132455295,
"grad_norm": 0.13848088681697845,
"learning_rate": 9.278062281171393e-05,
"loss": 0.2463,
"step": 523
},
{
"epoch": 0.056108791091123246,
"grad_norm": 0.19286379218101501,
"learning_rate": 9.246731944720675e-05,
"loss": 0.3067,
"step": 524
},
{
"epoch": 0.05621586893671699,
"grad_norm": 0.16500720381736755,
"learning_rate": 9.215409042721552e-05,
"loss": 0.3342,
"step": 525
},
{
"epoch": 0.05632294678231074,
"grad_norm": 0.223133385181427,
"learning_rate": 9.184093884318425e-05,
"loss": 0.5118,
"step": 526
},
{
"epoch": 0.05643002462790449,
"grad_norm": 0.16499963402748108,
"learning_rate": 9.152786778579267e-05,
"loss": 0.3624,
"step": 527
},
{
"epoch": 0.05653710247349823,
"grad_norm": 0.14940662682056427,
"learning_rate": 9.121488034492569e-05,
"loss": 0.307,
"step": 528
},
{
"epoch": 0.05664418031909198,
"grad_norm": 0.15789055824279785,
"learning_rate": 9.090197960964301e-05,
"loss": 0.3075,
"step": 529
},
{
"epoch": 0.05675125816468573,
"grad_norm": 0.16344444453716278,
"learning_rate": 9.058916866814858e-05,
"loss": 0.3722,
"step": 530
},
{
"epoch": 0.056858336010279474,
"grad_norm": 0.18045423924922943,
"learning_rate": 9.027645060776006e-05,
"loss": 0.4066,
"step": 531
},
{
"epoch": 0.05696541385587322,
"grad_norm": 0.14639213681221008,
"learning_rate": 8.99638285148785e-05,
"loss": 0.24,
"step": 532
},
{
"epoch": 0.057072491701466964,
"grad_norm": 0.1606295108795166,
"learning_rate": 8.965130547495776e-05,
"loss": 0.3495,
"step": 533
},
{
"epoch": 0.057179569547060716,
"grad_norm": 0.17020860314369202,
"learning_rate": 8.933888457247402e-05,
"loss": 0.3995,
"step": 534
},
{
"epoch": 0.05728664739265446,
"grad_norm": 0.16404946148395538,
"learning_rate": 8.902656889089548e-05,
"loss": 0.3087,
"step": 535
},
{
"epoch": 0.057393725238248205,
"grad_norm": 0.16885989904403687,
"learning_rate": 8.871436151265184e-05,
"loss": 0.3451,
"step": 536
},
{
"epoch": 0.05750080308384195,
"grad_norm": 0.16032719612121582,
"learning_rate": 8.840226551910387e-05,
"loss": 0.3151,
"step": 537
},
{
"epoch": 0.0576078809294357,
"grad_norm": 0.16192081570625305,
"learning_rate": 8.809028399051302e-05,
"loss": 0.3284,
"step": 538
},
{
"epoch": 0.05771495877502945,
"grad_norm": 0.16214251518249512,
"learning_rate": 8.777842000601105e-05,
"loss": 0.2159,
"step": 539
},
{
"epoch": 0.05782203662062319,
"grad_norm": 0.2248457819223404,
"learning_rate": 8.746667664356956e-05,
"loss": 0.3886,
"step": 540
},
{
"epoch": 0.05792911446621694,
"grad_norm": 0.17317667603492737,
"learning_rate": 8.715505697996971e-05,
"loss": 0.2789,
"step": 541
},
{
"epoch": 0.05803619231181069,
"grad_norm": 0.18882204592227936,
"learning_rate": 8.684356409077176e-05,
"loss": 0.3398,
"step": 542
},
{
"epoch": 0.058143270157404434,
"grad_norm": 0.2111193686723709,
"learning_rate": 8.653220105028474e-05,
"loss": 0.4288,
"step": 543
},
{
"epoch": 0.05825034800299818,
"grad_norm": 0.14698849618434906,
"learning_rate": 8.62209709315362e-05,
"loss": 0.3265,
"step": 544
},
{
"epoch": 0.05835742584859192,
"grad_norm": 0.15182967483997345,
"learning_rate": 8.590987680624174e-05,
"loss": 0.2549,
"step": 545
},
{
"epoch": 0.058464503694185675,
"grad_norm": 0.1783917099237442,
"learning_rate": 8.559892174477479e-05,
"loss": 0.4651,
"step": 546
},
{
"epoch": 0.05857158153977942,
"grad_norm": 0.2132243663072586,
"learning_rate": 8.528810881613626e-05,
"loss": 0.3032,
"step": 547
},
{
"epoch": 0.058678659385373165,
"grad_norm": 0.18143096566200256,
"learning_rate": 8.497744108792429e-05,
"loss": 0.2871,
"step": 548
},
{
"epoch": 0.05878573723096691,
"grad_norm": 0.1841660588979721,
"learning_rate": 8.466692162630392e-05,
"loss": 0.3478,
"step": 549
},
{
"epoch": 0.05889281507656066,
"grad_norm": 0.16989177465438843,
"learning_rate": 8.435655349597689e-05,
"loss": 0.3273,
"step": 550
},
{
"epoch": 0.05899989292215441,
"grad_norm": 0.1569872945547104,
"learning_rate": 8.404633976015134e-05,
"loss": 0.2909,
"step": 551
},
{
"epoch": 0.05910697076774815,
"grad_norm": 0.15851524472236633,
"learning_rate": 8.373628348051165e-05,
"loss": 0.2183,
"step": 552
},
{
"epoch": 0.059214048613341896,
"grad_norm": 0.16590699553489685,
"learning_rate": 8.342638771718802e-05,
"loss": 0.215,
"step": 553
},
{
"epoch": 0.05932112645893565,
"grad_norm": 0.2867710590362549,
"learning_rate": 8.311665552872662e-05,
"loss": 0.4176,
"step": 554
},
{
"epoch": 0.05942820430452939,
"grad_norm": 0.16975639760494232,
"learning_rate": 8.280708997205904e-05,
"loss": 0.34,
"step": 555
},
{
"epoch": 0.05953528215012314,
"grad_norm": 0.15163715183734894,
"learning_rate": 8.249769410247239e-05,
"loss": 0.275,
"step": 556
},
{
"epoch": 0.05964235999571688,
"grad_norm": 0.1733134686946869,
"learning_rate": 8.218847097357898e-05,
"loss": 0.2584,
"step": 557
},
{
"epoch": 0.059749437841310635,
"grad_norm": 0.16540873050689697,
"learning_rate": 8.187942363728625e-05,
"loss": 0.2495,
"step": 558
},
{
"epoch": 0.05985651568690438,
"grad_norm": 0.18063515424728394,
"learning_rate": 8.157055514376666e-05,
"loss": 0.3324,
"step": 559
},
{
"epoch": 0.059963593532498125,
"grad_norm": 0.1537630259990692,
"learning_rate": 8.126186854142752e-05,
"loss": 0.2322,
"step": 560
},
{
"epoch": 0.06007067137809187,
"grad_norm": 0.14401783049106598,
"learning_rate": 8.095336687688102e-05,
"loss": 0.2909,
"step": 561
},
{
"epoch": 0.06017774922368562,
"grad_norm": 0.19324550032615662,
"learning_rate": 8.064505319491398e-05,
"loss": 0.4552,
"step": 562
},
{
"epoch": 0.060284827069279366,
"grad_norm": 0.13768549263477325,
"learning_rate": 8.033693053845801e-05,
"loss": 0.2493,
"step": 563
},
{
"epoch": 0.06039190491487311,
"grad_norm": 0.17372079193592072,
"learning_rate": 8.002900194855932e-05,
"loss": 0.3261,
"step": 564
},
{
"epoch": 0.06049898276046686,
"grad_norm": 0.18107670545578003,
"learning_rate": 7.972127046434878e-05,
"loss": 0.3877,
"step": 565
},
{
"epoch": 0.06060606060606061,
"grad_norm": 0.1908927708864212,
"learning_rate": 7.941373912301189e-05,
"loss": 0.4814,
"step": 566
},
{
"epoch": 0.06071313845165435,
"grad_norm": 0.20689429342746735,
"learning_rate": 7.910641095975886e-05,
"loss": 0.432,
"step": 567
},
{
"epoch": 0.0608202162972481,
"grad_norm": 0.18664167821407318,
"learning_rate": 7.879928900779456e-05,
"loss": 0.3784,
"step": 568
},
{
"epoch": 0.06092729414284185,
"grad_norm": 0.16357581317424774,
"learning_rate": 7.849237629828869e-05,
"loss": 0.3061,
"step": 569
},
{
"epoch": 0.061034371988435594,
"grad_norm": 0.20243744552135468,
"learning_rate": 7.818567586034577e-05,
"loss": 0.326,
"step": 570
},
{
"epoch": 0.06114144983402934,
"grad_norm": 0.15626615285873413,
"learning_rate": 7.787919072097531e-05,
"loss": 0.3051,
"step": 571
},
{
"epoch": 0.061248527679623084,
"grad_norm": 0.16009844839572906,
"learning_rate": 7.75729239050619e-05,
"loss": 0.3191,
"step": 572
},
{
"epoch": 0.061355605525216836,
"grad_norm": 0.19337281584739685,
"learning_rate": 7.726687843533538e-05,
"loss": 0.4369,
"step": 573
},
{
"epoch": 0.06146268337081058,
"grad_norm": 0.17653042078018188,
"learning_rate": 7.696105733234098e-05,
"loss": 0.3309,
"step": 574
},
{
"epoch": 0.061569761216404326,
"grad_norm": 0.16593199968338013,
"learning_rate": 7.66554636144095e-05,
"loss": 0.44,
"step": 575
},
{
"epoch": 0.06167683906199807,
"grad_norm": 0.16403137147426605,
"learning_rate": 7.635010029762756e-05,
"loss": 0.2433,
"step": 576
},
{
"epoch": 0.06178391690759182,
"grad_norm": 0.13667182624340057,
"learning_rate": 7.604497039580785e-05,
"loss": 0.2799,
"step": 577
},
{
"epoch": 0.06189099475318557,
"grad_norm": 0.16916875541210175,
"learning_rate": 7.574007692045928e-05,
"loss": 0.3701,
"step": 578
},
{
"epoch": 0.06199807259877931,
"grad_norm": 0.16624587774276733,
"learning_rate": 7.543542288075739e-05,
"loss": 0.2409,
"step": 579
},
{
"epoch": 0.06210515044437306,
"grad_norm": 0.2018321007490158,
"learning_rate": 7.513101128351454e-05,
"loss": 0.3169,
"step": 580
},
{
"epoch": 0.06221222828996681,
"grad_norm": 0.16380462050437927,
"learning_rate": 7.48268451331503e-05,
"loss": 0.4424,
"step": 581
},
{
"epoch": 0.062319306135560554,
"grad_norm": 0.14434152841567993,
"learning_rate": 7.45229274316618e-05,
"loss": 0.2669,
"step": 582
},
{
"epoch": 0.0624263839811543,
"grad_norm": 0.22366313636302948,
"learning_rate": 7.421926117859403e-05,
"loss": 0.455,
"step": 583
},
{
"epoch": 0.06253346182674804,
"grad_norm": 0.18707706034183502,
"learning_rate": 7.391584937101033e-05,
"loss": 0.3651,
"step": 584
},
{
"epoch": 0.06264053967234179,
"grad_norm": 0.15041711926460266,
"learning_rate": 7.361269500346274e-05,
"loss": 0.3066,
"step": 585
},
{
"epoch": 0.06274761751793553,
"grad_norm": 0.19988040626049042,
"learning_rate": 7.330980106796246e-05,
"loss": 0.3106,
"step": 586
},
{
"epoch": 0.06285469536352929,
"grad_norm": 0.18236958980560303,
"learning_rate": 7.300717055395039e-05,
"loss": 0.3529,
"step": 587
},
{
"epoch": 0.06296177320912304,
"grad_norm": 0.15908972918987274,
"learning_rate": 7.270480644826749e-05,
"loss": 0.262,
"step": 588
},
{
"epoch": 0.06306885105471678,
"grad_norm": 0.16096021234989166,
"learning_rate": 7.240271173512546e-05,
"loss": 0.3349,
"step": 589
},
{
"epoch": 0.06317592890031053,
"grad_norm": 0.14579513669013977,
"learning_rate": 7.210088939607708e-05,
"loss": 0.3535,
"step": 590
},
{
"epoch": 0.06328300674590427,
"grad_norm": 0.1794016808271408,
"learning_rate": 7.179934240998706e-05,
"loss": 0.3372,
"step": 591
},
{
"epoch": 0.06339008459149802,
"grad_norm": 0.1510874330997467,
"learning_rate": 7.149807375300239e-05,
"loss": 0.2272,
"step": 592
},
{
"epoch": 0.06349716243709176,
"grad_norm": 0.23262274265289307,
"learning_rate": 7.119708639852312e-05,
"loss": 0.4265,
"step": 593
},
{
"epoch": 0.0636042402826855,
"grad_norm": 0.20220021903514862,
"learning_rate": 7.089638331717284e-05,
"loss": 0.4163,
"step": 594
},
{
"epoch": 0.06371131812827927,
"grad_norm": 0.1835254728794098,
"learning_rate": 7.059596747676962e-05,
"loss": 0.3373,
"step": 595
},
{
"epoch": 0.06381839597387301,
"grad_norm": 0.1331084668636322,
"learning_rate": 7.029584184229653e-05,
"loss": 0.2568,
"step": 596
},
{
"epoch": 0.06392547381946675,
"grad_norm": 0.1771346926689148,
"learning_rate": 6.999600937587239e-05,
"loss": 0.3663,
"step": 597
},
{
"epoch": 0.0640325516650605,
"grad_norm": 0.16743473708629608,
"learning_rate": 6.969647303672262e-05,
"loss": 0.292,
"step": 598
},
{
"epoch": 0.06413962951065424,
"grad_norm": 0.15180084109306335,
"learning_rate": 6.939723578114993e-05,
"loss": 0.2849,
"step": 599
},
{
"epoch": 0.06424670735624799,
"grad_norm": 0.15079252421855927,
"learning_rate": 6.909830056250527e-05,
"loss": 0.3499,
"step": 600
},
{
"epoch": 0.06435378520184173,
"grad_norm": 0.16857562959194183,
"learning_rate": 6.879967033115853e-05,
"loss": 0.3369,
"step": 601
},
{
"epoch": 0.06446086304743548,
"grad_norm": 0.15610823035240173,
"learning_rate": 6.850134803446954e-05,
"loss": 0.4078,
"step": 602
},
{
"epoch": 0.06456794089302924,
"grad_norm": 0.16949012875556946,
"learning_rate": 6.820333661675893e-05,
"loss": 0.326,
"step": 603
},
{
"epoch": 0.06467501873862298,
"grad_norm": 0.15723907947540283,
"learning_rate": 6.790563901927907e-05,
"loss": 0.3191,
"step": 604
},
{
"epoch": 0.06478209658421673,
"grad_norm": 0.1530887633562088,
"learning_rate": 6.760825818018508e-05,
"loss": 0.3864,
"step": 605
},
{
"epoch": 0.06488917442981047,
"grad_norm": 0.1763346791267395,
"learning_rate": 6.731119703450577e-05,
"loss": 0.3351,
"step": 606
},
{
"epoch": 0.06499625227540422,
"grad_norm": 0.1836361289024353,
"learning_rate": 6.701445851411472e-05,
"loss": 0.307,
"step": 607
},
{
"epoch": 0.06510333012099796,
"grad_norm": 0.1673685908317566,
"learning_rate": 6.671804554770135e-05,
"loss": 0.2753,
"step": 608
},
{
"epoch": 0.06521040796659171,
"grad_norm": 0.1554986834526062,
"learning_rate": 6.642196106074194e-05,
"loss": 0.3076,
"step": 609
},
{
"epoch": 0.06531748581218545,
"grad_norm": 0.15603835880756378,
"learning_rate": 6.612620797547087e-05,
"loss": 0.3019,
"step": 610
},
{
"epoch": 0.06542456365777921,
"grad_norm": 0.1522570699453354,
"learning_rate": 6.583078921085167e-05,
"loss": 0.2793,
"step": 611
},
{
"epoch": 0.06553164150337296,
"grad_norm": 0.18546074628829956,
"learning_rate": 6.55357076825483e-05,
"loss": 0.3955,
"step": 612
},
{
"epoch": 0.0656387193489667,
"grad_norm": 0.14978669583797455,
"learning_rate": 6.52409663028963e-05,
"loss": 0.3413,
"step": 613
},
{
"epoch": 0.06574579719456045,
"grad_norm": 0.2138916552066803,
"learning_rate": 6.494656798087412e-05,
"loss": 0.3169,
"step": 614
},
{
"epoch": 0.06585287504015419,
"grad_norm": 0.1433931142091751,
"learning_rate": 6.465251562207431e-05,
"loss": 0.2771,
"step": 615
},
{
"epoch": 0.06595995288574794,
"grad_norm": 0.1670522540807724,
"learning_rate": 6.435881212867493e-05,
"loss": 0.3817,
"step": 616
},
{
"epoch": 0.06606703073134168,
"grad_norm": 0.17601914703845978,
"learning_rate": 6.406546039941094e-05,
"loss": 0.3462,
"step": 617
},
{
"epoch": 0.06617410857693543,
"grad_norm": 0.1717825084924698,
"learning_rate": 6.377246332954544e-05,
"loss": 0.328,
"step": 618
},
{
"epoch": 0.06628118642252918,
"grad_norm": 0.16034941375255585,
"learning_rate": 6.347982381084123e-05,
"loss": 0.2962,
"step": 619
},
{
"epoch": 0.06638826426812293,
"grad_norm": 0.1435386687517166,
"learning_rate": 6.318754473153221e-05,
"loss": 0.2272,
"step": 620
},
{
"epoch": 0.06649534211371667,
"grad_norm": 0.13801023364067078,
"learning_rate": 6.289562897629492e-05,
"loss": 0.2799,
"step": 621
},
{
"epoch": 0.06660241995931042,
"grad_norm": 0.15614642202854156,
"learning_rate": 6.260407942621998e-05,
"loss": 0.441,
"step": 622
},
{
"epoch": 0.06670949780490416,
"grad_norm": 0.16773013770580292,
"learning_rate": 6.231289895878375e-05,
"loss": 0.2803,
"step": 623
},
{
"epoch": 0.06681657565049791,
"grad_norm": 0.16002804040908813,
"learning_rate": 6.20220904478199e-05,
"loss": 0.3051,
"step": 624
},
{
"epoch": 0.06692365349609165,
"grad_norm": 0.15663942694664001,
"learning_rate": 6.173165676349103e-05,
"loss": 0.3281,
"step": 625
},
{
"epoch": 0.0670307313416854,
"grad_norm": 0.1416623592376709,
"learning_rate": 6.144160077226036e-05,
"loss": 0.2683,
"step": 626
},
{
"epoch": 0.06713780918727916,
"grad_norm": 0.15320400893688202,
"learning_rate": 6.11519253368634e-05,
"loss": 0.2714,
"step": 627
},
{
"epoch": 0.0672448870328729,
"grad_norm": 0.19521519541740417,
"learning_rate": 6.086263331627976e-05,
"loss": 0.5039,
"step": 628
},
{
"epoch": 0.06735196487846665,
"grad_norm": 0.182010218501091,
"learning_rate": 6.05737275657049e-05,
"loss": 0.4331,
"step": 629
},
{
"epoch": 0.06745904272406039,
"grad_norm": 0.20210574567317963,
"learning_rate": 6.0285210936521955e-05,
"loss": 0.3858,
"step": 630
},
{
"epoch": 0.06756612056965414,
"grad_norm": 0.15948592126369476,
"learning_rate": 5.999708627627354e-05,
"loss": 0.2308,
"step": 631
},
{
"epoch": 0.06767319841524788,
"grad_norm": 0.16815067827701569,
"learning_rate": 5.9709356428633746e-05,
"loss": 0.2901,
"step": 632
},
{
"epoch": 0.06778027626084163,
"grad_norm": 0.18218070268630981,
"learning_rate": 5.9422024233380013e-05,
"loss": 0.3229,
"step": 633
},
{
"epoch": 0.06788735410643539,
"grad_norm": 0.1760886311531067,
"learning_rate": 5.913509252636511e-05,
"loss": 0.2355,
"step": 634
},
{
"epoch": 0.06799443195202913,
"grad_norm": 0.14710214734077454,
"learning_rate": 5.884856413948913e-05,
"loss": 0.3329,
"step": 635
},
{
"epoch": 0.06810150979762288,
"grad_norm": 0.1451999992132187,
"learning_rate": 5.856244190067159e-05,
"loss": 0.2894,
"step": 636
},
{
"epoch": 0.06820858764321662,
"grad_norm": 0.17219161987304688,
"learning_rate": 5.82767286338235e-05,
"loss": 0.3896,
"step": 637
},
{
"epoch": 0.06831566548881036,
"grad_norm": 0.17436614632606506,
"learning_rate": 5.799142715881938e-05,
"loss": 0.4378,
"step": 638
},
{
"epoch": 0.06842274333440411,
"grad_norm": 0.13370436429977417,
"learning_rate": 5.770654029146969e-05,
"loss": 0.2642,
"step": 639
},
{
"epoch": 0.06852982117999785,
"grad_norm": 0.1600145697593689,
"learning_rate": 5.7422070843492734e-05,
"loss": 0.2838,
"step": 640
},
{
"epoch": 0.0686368990255916,
"grad_norm": 0.18459884822368622,
"learning_rate": 5.713802162248718e-05,
"loss": 0.3208,
"step": 641
},
{
"epoch": 0.06874397687118536,
"grad_norm": 0.13782131671905518,
"learning_rate": 5.6854395431904094e-05,
"loss": 0.3096,
"step": 642
},
{
"epoch": 0.0688510547167791,
"grad_norm": 0.1642604023218155,
"learning_rate": 5.657119507101954e-05,
"loss": 0.4043,
"step": 643
},
{
"epoch": 0.06895813256237285,
"grad_norm": 0.17711035907268524,
"learning_rate": 5.6288423334906735e-05,
"loss": 0.3593,
"step": 644
},
{
"epoch": 0.06906521040796659,
"grad_norm": 0.17758657038211823,
"learning_rate": 5.6006083014408484e-05,
"loss": 0.3003,
"step": 645
},
{
"epoch": 0.06917228825356034,
"grad_norm": 0.18889419734477997,
"learning_rate": 5.572417689610987e-05,
"loss": 0.3131,
"step": 646
},
{
"epoch": 0.06927936609915408,
"grad_norm": 0.20904329419136047,
"learning_rate": 5.544270776231038e-05,
"loss": 0.3291,
"step": 647
},
{
"epoch": 0.06938644394474783,
"grad_norm": 0.15047985315322876,
"learning_rate": 5.5161678390996796e-05,
"loss": 0.3102,
"step": 648
},
{
"epoch": 0.06949352179034157,
"grad_norm": 0.23132561147212982,
"learning_rate": 5.488109155581549e-05,
"loss": 0.4353,
"step": 649
},
{
"epoch": 0.06960059963593533,
"grad_norm": 0.19946032762527466,
"learning_rate": 5.4600950026045326e-05,
"loss": 0.3372,
"step": 650
},
{
"epoch": 0.06970767748152908,
"grad_norm": 0.20680102705955505,
"learning_rate": 5.4321256566570036e-05,
"loss": 0.4628,
"step": 651
},
{
"epoch": 0.06981475532712282,
"grad_norm": 0.15579630434513092,
"learning_rate": 5.404201393785122e-05,
"loss": 0.2367,
"step": 652
},
{
"epoch": 0.06992183317271657,
"grad_norm": 0.20350880920886993,
"learning_rate": 5.3763224895900846e-05,
"loss": 0.3568,
"step": 653
},
{
"epoch": 0.07002891101831031,
"grad_norm": 0.17040583491325378,
"learning_rate": 5.348489219225416e-05,
"loss": 0.3297,
"step": 654
},
{
"epoch": 0.07013598886390406,
"grad_norm": 0.22072608768939972,
"learning_rate": 5.320701857394268e-05,
"loss": 0.3243,
"step": 655
},
{
"epoch": 0.0702430667094978,
"grad_norm": 0.16529580950737,
"learning_rate": 5.292960678346675e-05,
"loss": 0.3504,
"step": 656
},
{
"epoch": 0.07035014455509155,
"grad_norm": 0.17127272486686707,
"learning_rate": 5.265265955876879e-05,
"loss": 0.3472,
"step": 657
},
{
"epoch": 0.0704572224006853,
"grad_norm": 0.18526853621006012,
"learning_rate": 5.237617963320608e-05,
"loss": 0.302,
"step": 658
},
{
"epoch": 0.07056430024627905,
"grad_norm": 0.18870258331298828,
"learning_rate": 5.210016973552391e-05,
"loss": 0.3709,
"step": 659
},
{
"epoch": 0.0706713780918728,
"grad_norm": 0.1274450570344925,
"learning_rate": 5.182463258982846e-05,
"loss": 0.2597,
"step": 660
},
{
"epoch": 0.07077845593746654,
"grad_norm": 0.17223431169986725,
"learning_rate": 5.1549570915560206e-05,
"loss": 0.3575,
"step": 661
},
{
"epoch": 0.07088553378306028,
"grad_norm": 0.16141627728939056,
"learning_rate": 5.127498742746675e-05,
"loss": 0.3195,
"step": 662
},
{
"epoch": 0.07099261162865403,
"grad_norm": 0.24255283176898956,
"learning_rate": 5.100088483557634e-05,
"loss": 0.4154,
"step": 663
},
{
"epoch": 0.07109968947424777,
"grad_norm": 0.18947787582874298,
"learning_rate": 5.072726584517086e-05,
"loss": 0.3083,
"step": 664
},
{
"epoch": 0.07120676731984152,
"grad_norm": 0.16710755228996277,
"learning_rate": 5.045413315675924e-05,
"loss": 0.3242,
"step": 665
},
{
"epoch": 0.07131384516543528,
"grad_norm": 0.16929441690444946,
"learning_rate": 5.018148946605092e-05,
"loss": 0.2543,
"step": 666
},
{
"epoch": 0.07142092301102902,
"grad_norm": 0.15913401544094086,
"learning_rate": 4.990933746392899e-05,
"loss": 0.3979,
"step": 667
},
{
"epoch": 0.07152800085662277,
"grad_norm": 0.2128509283065796,
"learning_rate": 4.9637679836423924e-05,
"loss": 0.3802,
"step": 668
},
{
"epoch": 0.07163507870221651,
"grad_norm": 0.19346699118614197,
"learning_rate": 4.9366519264686725e-05,
"loss": 0.3781,
"step": 669
},
{
"epoch": 0.07174215654781026,
"grad_norm": 0.19327999651432037,
"learning_rate": 4.909585842496287e-05,
"loss": 0.3583,
"step": 670
},
{
"epoch": 0.071849234393404,
"grad_norm": 0.13450753688812256,
"learning_rate": 4.8825699988565485e-05,
"loss": 0.2058,
"step": 671
},
{
"epoch": 0.07195631223899775,
"grad_norm": 0.16594289243221283,
"learning_rate": 4.8556046621849346e-05,
"loss": 0.2914,
"step": 672
},
{
"epoch": 0.07206339008459149,
"grad_norm": 0.17473144829273224,
"learning_rate": 4.828690098618429e-05,
"loss": 0.3397,
"step": 673
},
{
"epoch": 0.07217046793018525,
"grad_norm": 0.1864292323589325,
"learning_rate": 4.8018265737929044e-05,
"loss": 0.4704,
"step": 674
},
{
"epoch": 0.072277545775779,
"grad_norm": 0.17810168862342834,
"learning_rate": 4.7750143528405126e-05,
"loss": 0.3505,
"step": 675
},
{
"epoch": 0.07238462362137274,
"grad_norm": 0.15457181632518768,
"learning_rate": 4.748253700387042e-05,
"loss": 0.323,
"step": 676
},
{
"epoch": 0.07249170146696648,
"grad_norm": 0.1830655038356781,
"learning_rate": 4.721544880549337e-05,
"loss": 0.332,
"step": 677
},
{
"epoch": 0.07259877931256023,
"grad_norm": 0.1661139875650406,
"learning_rate": 4.694888156932658e-05,
"loss": 0.2828,
"step": 678
},
{
"epoch": 0.07270585715815397,
"grad_norm": 0.1575031727552414,
"learning_rate": 4.668283792628114e-05,
"loss": 0.3252,
"step": 679
},
{
"epoch": 0.07281293500374772,
"grad_norm": 0.1568552553653717,
"learning_rate": 4.6417320502100316e-05,
"loss": 0.2818,
"step": 680
},
{
"epoch": 0.07292001284934146,
"grad_norm": 0.18243847787380219,
"learning_rate": 4.615233191733398e-05,
"loss": 0.3775,
"step": 681
},
{
"epoch": 0.07302709069493522,
"grad_norm": 0.14313600957393646,
"learning_rate": 4.588787478731242e-05,
"loss": 0.2801,
"step": 682
},
{
"epoch": 0.07313416854052897,
"grad_norm": 0.2518500089645386,
"learning_rate": 4.5623951722120736e-05,
"loss": 0.3921,
"step": 683
},
{
"epoch": 0.07324124638612271,
"grad_norm": 0.1808672994375229,
"learning_rate": 4.5360565326573104e-05,
"loss": 0.3468,
"step": 684
},
{
"epoch": 0.07334832423171646,
"grad_norm": 0.1666547656059265,
"learning_rate": 4.5097718200186814e-05,
"loss": 0.3815,
"step": 685
},
{
"epoch": 0.0734554020773102,
"grad_norm": 0.1575447916984558,
"learning_rate": 4.483541293715698e-05,
"loss": 0.3646,
"step": 686
},
{
"epoch": 0.07356247992290395,
"grad_norm": 0.17928656935691833,
"learning_rate": 4.457365212633058e-05,
"loss": 0.3603,
"step": 687
},
{
"epoch": 0.07366955776849769,
"grad_norm": 0.19365298748016357,
"learning_rate": 4.431243835118124e-05,
"loss": 0.367,
"step": 688
},
{
"epoch": 0.07377663561409144,
"grad_norm": 0.1505446434020996,
"learning_rate": 4.4051774189783315e-05,
"loss": 0.2705,
"step": 689
},
{
"epoch": 0.0738837134596852,
"grad_norm": 0.18016034364700317,
"learning_rate": 4.379166221478697e-05,
"loss": 0.2904,
"step": 690
},
{
"epoch": 0.07399079130527894,
"grad_norm": 0.14219722151756287,
"learning_rate": 4.3532104993392306e-05,
"loss": 0.2767,
"step": 691
},
{
"epoch": 0.07409786915087269,
"grad_norm": 0.2010519653558731,
"learning_rate": 4.327310508732437e-05,
"loss": 0.4859,
"step": 692
},
{
"epoch": 0.07420494699646643,
"grad_norm": 0.15180367231369019,
"learning_rate": 4.301466505280762e-05,
"loss": 0.3175,
"step": 693
},
{
"epoch": 0.07431202484206018,
"grad_norm": 0.16391538083553314,
"learning_rate": 4.2756787440540936e-05,
"loss": 0.3576,
"step": 694
},
{
"epoch": 0.07441910268765392,
"grad_norm": 0.1550375074148178,
"learning_rate": 4.249947479567218e-05,
"loss": 0.3563,
"step": 695
},
{
"epoch": 0.07452618053324767,
"grad_norm": 0.16695566475391388,
"learning_rate": 4.224272965777326e-05,
"loss": 0.2697,
"step": 696
},
{
"epoch": 0.07463325837884142,
"grad_norm": 0.16913089156150818,
"learning_rate": 4.1986554560815096e-05,
"loss": 0.3506,
"step": 697
},
{
"epoch": 0.07474033622443517,
"grad_norm": 0.15745219588279724,
"learning_rate": 4.173095203314241e-05,
"loss": 0.311,
"step": 698
},
{
"epoch": 0.07484741407002891,
"grad_norm": 0.1870749592781067,
"learning_rate": 4.1475924597449024e-05,
"loss": 0.3289,
"step": 699
},
{
"epoch": 0.07495449191562266,
"grad_norm": 0.18129318952560425,
"learning_rate": 4.12214747707527e-05,
"loss": 0.3735,
"step": 700
},
{
"epoch": 0.0750615697612164,
"grad_norm": 0.1702040284872055,
"learning_rate": 4.096760506437057e-05,
"loss": 0.3375,
"step": 701
},
{
"epoch": 0.07516864760681015,
"grad_norm": 0.16838037967681885,
"learning_rate": 4.071431798389408e-05,
"loss": 0.3685,
"step": 702
},
{
"epoch": 0.0752757254524039,
"grad_norm": 0.22011017799377441,
"learning_rate": 4.0461616029164526e-05,
"loss": 0.3469,
"step": 703
},
{
"epoch": 0.07538280329799764,
"grad_norm": 0.14272677898406982,
"learning_rate": 4.020950169424815e-05,
"loss": 0.3497,
"step": 704
},
{
"epoch": 0.0754898811435914,
"grad_norm": 0.15662071108818054,
"learning_rate": 3.9957977467411615e-05,
"loss": 0.2359,
"step": 705
},
{
"epoch": 0.07559695898918514,
"grad_norm": 0.12998908758163452,
"learning_rate": 3.9707045831097555e-05,
"loss": 0.2384,
"step": 706
},
{
"epoch": 0.07570403683477889,
"grad_norm": 0.18728865683078766,
"learning_rate": 3.945670926189987e-05,
"loss": 0.3876,
"step": 707
},
{
"epoch": 0.07581111468037263,
"grad_norm": 0.19201086461544037,
"learning_rate": 3.920697023053949e-05,
"loss": 0.4129,
"step": 708
},
{
"epoch": 0.07591819252596638,
"grad_norm": 0.15770968794822693,
"learning_rate": 3.895783120183976e-05,
"loss": 0.3428,
"step": 709
},
{
"epoch": 0.07602527037156012,
"grad_norm": 0.16312432289123535,
"learning_rate": 3.8709294634702376e-05,
"loss": 0.3468,
"step": 710
},
{
"epoch": 0.07613234821715387,
"grad_norm": 0.15018045902252197,
"learning_rate": 3.846136298208285e-05,
"loss": 0.2924,
"step": 711
},
{
"epoch": 0.07623942606274761,
"grad_norm": 0.150425523519516,
"learning_rate": 3.821403869096658e-05,
"loss": 0.2774,
"step": 712
},
{
"epoch": 0.07634650390834137,
"grad_norm": 0.2021748274564743,
"learning_rate": 3.796732420234443e-05,
"loss": 0.3377,
"step": 713
},
{
"epoch": 0.07645358175393512,
"grad_norm": 0.18799428641796112,
"learning_rate": 3.7721221951188765e-05,
"loss": 0.3719,
"step": 714
},
{
"epoch": 0.07656065959952886,
"grad_norm": 0.1765313595533371,
"learning_rate": 3.747573436642951e-05,
"loss": 0.3921,
"step": 715
},
{
"epoch": 0.0766677374451226,
"grad_norm": 0.15740084648132324,
"learning_rate": 3.7230863870929964e-05,
"loss": 0.3102,
"step": 716
},
{
"epoch": 0.07677481529071635,
"grad_norm": 0.17501920461654663,
"learning_rate": 3.698661288146311e-05,
"loss": 0.4363,
"step": 717
},
{
"epoch": 0.0768818931363101,
"grad_norm": 0.2245578169822693,
"learning_rate": 3.674298380868756e-05,
"loss": 0.5983,
"step": 718
},
{
"epoch": 0.07698897098190384,
"grad_norm": 0.18928749859333038,
"learning_rate": 3.649997905712396e-05,
"loss": 0.2717,
"step": 719
},
{
"epoch": 0.07709604882749758,
"grad_norm": 0.14683455228805542,
"learning_rate": 3.6257601025131026e-05,
"loss": 0.2873,
"step": 720
},
{
"epoch": 0.07720312667309134,
"grad_norm": 0.19283850491046906,
"learning_rate": 3.601585210488218e-05,
"loss": 0.2882,
"step": 721
},
{
"epoch": 0.07731020451868509,
"grad_norm": 0.21515490114688873,
"learning_rate": 3.577473468234156e-05,
"loss": 0.3484,
"step": 722
},
{
"epoch": 0.07741728236427883,
"grad_norm": 0.16853037476539612,
"learning_rate": 3.553425113724088e-05,
"loss": 0.3759,
"step": 723
},
{
"epoch": 0.07752436020987258,
"grad_norm": 0.1734563559293747,
"learning_rate": 3.52944038430556e-05,
"loss": 0.3769,
"step": 724
},
{
"epoch": 0.07763143805546632,
"grad_norm": 0.1489877551794052,
"learning_rate": 3.5055195166981645e-05,
"loss": 0.3306,
"step": 725
},
{
"epoch": 0.07773851590106007,
"grad_norm": 0.1704966127872467,
"learning_rate": 3.481662746991214e-05,
"loss": 0.3635,
"step": 726
},
{
"epoch": 0.07784559374665381,
"grad_norm": 0.13263505697250366,
"learning_rate": 3.4578703106413904e-05,
"loss": 0.2811,
"step": 727
},
{
"epoch": 0.07795267159224756,
"grad_norm": 0.1653466671705246,
"learning_rate": 3.4341424424704375e-05,
"loss": 0.301,
"step": 728
},
{
"epoch": 0.07805974943784132,
"grad_norm": 0.17807163298130035,
"learning_rate": 3.4104793766628304e-05,
"loss": 0.3877,
"step": 729
},
{
"epoch": 0.07816682728343506,
"grad_norm": 0.1522621512413025,
"learning_rate": 3.386881346763483e-05,
"loss": 0.2858,
"step": 730
},
{
"epoch": 0.0782739051290288,
"grad_norm": 0.15765105187892914,
"learning_rate": 3.363348585675414e-05,
"loss": 0.3006,
"step": 731
},
{
"epoch": 0.07838098297462255,
"grad_norm": 0.16139338910579681,
"learning_rate": 3.339881325657484e-05,
"loss": 0.3796,
"step": 732
},
{
"epoch": 0.0784880608202163,
"grad_norm": 0.15585535764694214,
"learning_rate": 3.316479798322072e-05,
"loss": 0.3781,
"step": 733
},
{
"epoch": 0.07859513866581004,
"grad_norm": 0.14251768589019775,
"learning_rate": 3.2931442346328004e-05,
"loss": 0.2142,
"step": 734
},
{
"epoch": 0.07870221651140379,
"grad_norm": 0.14421778917312622,
"learning_rate": 3.269874864902269e-05,
"loss": 0.282,
"step": 735
},
{
"epoch": 0.07880929435699753,
"grad_norm": 0.1328769028186798,
"learning_rate": 3.246671918789755e-05,
"loss": 0.2556,
"step": 736
},
{
"epoch": 0.07891637220259129,
"grad_norm": 0.1606833040714264,
"learning_rate": 3.223535625298979e-05,
"loss": 0.2804,
"step": 737
},
{
"epoch": 0.07902345004818503,
"grad_norm": 0.1721002757549286,
"learning_rate": 3.200466212775808e-05,
"loss": 0.2562,
"step": 738
},
{
"epoch": 0.07913052789377878,
"grad_norm": 0.1502068191766739,
"learning_rate": 3.1774639089060363e-05,
"loss": 0.319,
"step": 739
},
{
"epoch": 0.07923760573937252,
"grad_norm": 0.20539775490760803,
"learning_rate": 3.154528940713113e-05,
"loss": 0.3627,
"step": 740
},
{
"epoch": 0.07934468358496627,
"grad_norm": 0.20355840027332306,
"learning_rate": 3.1316615345559185e-05,
"loss": 0.3651,
"step": 741
},
{
"epoch": 0.07945176143056001,
"grad_norm": 0.18047836422920227,
"learning_rate": 3.108861916126518e-05,
"loss": 0.4033,
"step": 742
},
{
"epoch": 0.07955883927615376,
"grad_norm": 0.17290200293064117,
"learning_rate": 3.086130310447937e-05,
"loss": 0.3729,
"step": 743
},
{
"epoch": 0.0796659171217475,
"grad_norm": 0.1400056779384613,
"learning_rate": 3.063466941871952e-05,
"loss": 0.2791,
"step": 744
},
{
"epoch": 0.07977299496734126,
"grad_norm": 0.15376096963882446,
"learning_rate": 3.0408720340768572e-05,
"loss": 0.3535,
"step": 745
},
{
"epoch": 0.07988007281293501,
"grad_norm": 0.13626393675804138,
"learning_rate": 3.018345810065275e-05,
"loss": 0.2482,
"step": 746
},
{
"epoch": 0.07998715065852875,
"grad_norm": 0.17024940252304077,
"learning_rate": 2.9958884921619367e-05,
"loss": 0.3593,
"step": 747
},
{
"epoch": 0.0800942285041225,
"grad_norm": 0.20395411550998688,
"learning_rate": 2.9735003020115092e-05,
"loss": 0.4827,
"step": 748
},
{
"epoch": 0.08020130634971624,
"grad_norm": 0.18541432917118073,
"learning_rate": 2.9511814605763855e-05,
"loss": 0.3223,
"step": 749
},
{
"epoch": 0.08030838419530999,
"grad_norm": 0.1652992069721222,
"learning_rate": 2.9289321881345254e-05,
"loss": 0.3128,
"step": 750
},
{
"epoch": 0.08041546204090373,
"grad_norm": 0.19549131393432617,
"learning_rate": 2.9067527042772636e-05,
"loss": 0.4826,
"step": 751
},
{
"epoch": 0.08052253988649748,
"grad_norm": 0.15394878387451172,
"learning_rate": 2.8846432279071467e-05,
"loss": 0.3783,
"step": 752
},
{
"epoch": 0.08062961773209124,
"grad_norm": 0.15814286470413208,
"learning_rate": 2.8626039772357882e-05,
"loss": 0.324,
"step": 753
},
{
"epoch": 0.08073669557768498,
"grad_norm": 0.21982432901859283,
"learning_rate": 2.840635169781688e-05,
"loss": 0.45,
"step": 754
},
{
"epoch": 0.08084377342327873,
"grad_norm": 0.16086405515670776,
"learning_rate": 2.8187370223681132e-05,
"loss": 0.3717,
"step": 755
},
{
"epoch": 0.08095085126887247,
"grad_norm": 0.13305172324180603,
"learning_rate": 2.7969097511209308e-05,
"loss": 0.2966,
"step": 756
},
{
"epoch": 0.08105792911446622,
"grad_norm": 0.1707354336977005,
"learning_rate": 2.775153571466502e-05,
"loss": 0.2911,
"step": 757
},
{
"epoch": 0.08116500696005996,
"grad_norm": 0.14525282382965088,
"learning_rate": 2.753468698129533e-05,
"loss": 0.2842,
"step": 758
},
{
"epoch": 0.0812720848056537,
"grad_norm": 0.17985416948795319,
"learning_rate": 2.7318553451309726e-05,
"loss": 0.3126,
"step": 759
},
{
"epoch": 0.08137916265124746,
"grad_norm": 0.16003213822841644,
"learning_rate": 2.7103137257858868e-05,
"loss": 0.3547,
"step": 760
},
{
"epoch": 0.08148624049684121,
"grad_norm": 0.1999395340681076,
"learning_rate": 2.688844052701359e-05,
"loss": 0.348,
"step": 761
},
{
"epoch": 0.08159331834243495,
"grad_norm": 0.14586856961250305,
"learning_rate": 2.6674465377744017e-05,
"loss": 0.3054,
"step": 762
},
{
"epoch": 0.0817003961880287,
"grad_norm": 0.1372138261795044,
"learning_rate": 2.646121392189841e-05,
"loss": 0.2372,
"step": 763
},
{
"epoch": 0.08180747403362244,
"grad_norm": 0.15203608572483063,
"learning_rate": 2.624868826418262e-05,
"loss": 0.2984,
"step": 764
},
{
"epoch": 0.08191455187921619,
"grad_norm": 0.18717195093631744,
"learning_rate": 2.603689050213902e-05,
"loss": 0.2993,
"step": 765
},
{
"epoch": 0.08202162972480993,
"grad_norm": 0.17310731112957,
"learning_rate": 2.582582272612609e-05,
"loss": 0.2612,
"step": 766
},
{
"epoch": 0.08212870757040368,
"grad_norm": 0.1539638340473175,
"learning_rate": 2.561548701929749e-05,
"loss": 0.3713,
"step": 767
},
{
"epoch": 0.08223578541599744,
"grad_norm": 0.16853591799736023,
"learning_rate": 2.540588545758179e-05,
"loss": 0.2724,
"step": 768
},
{
"epoch": 0.08234286326159118,
"grad_norm": 0.1422484815120697,
"learning_rate": 2.5197020109661772e-05,
"loss": 0.273,
"step": 769
},
{
"epoch": 0.08244994110718493,
"grad_norm": 0.16871221363544464,
"learning_rate": 2.4988893036954043e-05,
"loss": 0.3818,
"step": 770
},
{
"epoch": 0.08255701895277867,
"grad_norm": 0.17967580258846283,
"learning_rate": 2.4781506293588873e-05,
"loss": 0.3771,
"step": 771
},
{
"epoch": 0.08266409679837242,
"grad_norm": 0.18379069864749908,
"learning_rate": 2.4574861926389615e-05,
"loss": 0.2425,
"step": 772
},
{
"epoch": 0.08277117464396616,
"grad_norm": 0.17387156188488007,
"learning_rate": 2.436896197485282e-05,
"loss": 0.4478,
"step": 773
},
{
"epoch": 0.0828782524895599,
"grad_norm": 0.19421961903572083,
"learning_rate": 2.4163808471127812e-05,
"loss": 0.3658,
"step": 774
},
{
"epoch": 0.08298533033515365,
"grad_norm": 0.14912578463554382,
"learning_rate": 2.3959403439996907e-05,
"loss": 0.338,
"step": 775
},
{
"epoch": 0.08309240818074741,
"grad_norm": 0.17342334985733032,
"learning_rate": 2.37557488988552e-05,
"loss": 0.3367,
"step": 776
},
{
"epoch": 0.08319948602634115,
"grad_norm": 0.1985640674829483,
"learning_rate": 2.3552846857690846e-05,
"loss": 0.3553,
"step": 777
},
{
"epoch": 0.0833065638719349,
"grad_norm": 0.16898934543132782,
"learning_rate": 2.3350699319065026e-05,
"loss": 0.3761,
"step": 778
},
{
"epoch": 0.08341364171752864,
"grad_norm": 0.17243239283561707,
"learning_rate": 2.3149308278092342e-05,
"loss": 0.3151,
"step": 779
},
{
"epoch": 0.08352071956312239,
"grad_norm": 0.19134309887886047,
"learning_rate": 2.2948675722421086e-05,
"loss": 0.3636,
"step": 780
},
{
"epoch": 0.08362779740871613,
"grad_norm": 0.20325163006782532,
"learning_rate": 2.2748803632213557e-05,
"loss": 0.3778,
"step": 781
},
{
"epoch": 0.08373487525430988,
"grad_norm": 0.14077632129192352,
"learning_rate": 2.254969398012663e-05,
"loss": 0.2707,
"step": 782
},
{
"epoch": 0.08384195309990362,
"grad_norm": 0.16174592077732086,
"learning_rate": 2.235134873129213e-05,
"loss": 0.3096,
"step": 783
},
{
"epoch": 0.08394903094549738,
"grad_norm": 0.1870279163122177,
"learning_rate": 2.2153769843297667e-05,
"loss": 0.3855,
"step": 784
},
{
"epoch": 0.08405610879109113,
"grad_norm": 0.1512727290391922,
"learning_rate": 2.195695926616702e-05,
"loss": 0.2219,
"step": 785
},
{
"epoch": 0.08416318663668487,
"grad_norm": 0.18673193454742432,
"learning_rate": 2.1760918942341192e-05,
"loss": 0.4747,
"step": 786
},
{
"epoch": 0.08427026448227862,
"grad_norm": 0.17171938717365265,
"learning_rate": 2.1565650806658975e-05,
"loss": 0.3874,
"step": 787
},
{
"epoch": 0.08437734232787236,
"grad_norm": 0.1813274323940277,
"learning_rate": 2.137115678633811e-05,
"loss": 0.375,
"step": 788
},
{
"epoch": 0.08448442017346611,
"grad_norm": 0.1640833467245102,
"learning_rate": 2.1177438800956007e-05,
"loss": 0.3012,
"step": 789
},
{
"epoch": 0.08459149801905985,
"grad_norm": 0.1751677691936493,
"learning_rate": 2.098449876243096e-05,
"loss": 0.3478,
"step": 790
},
{
"epoch": 0.0846985758646536,
"grad_norm": 0.14867408573627472,
"learning_rate": 2.07923385750033e-05,
"loss": 0.2817,
"step": 791
},
{
"epoch": 0.08480565371024736,
"grad_norm": 0.1769115924835205,
"learning_rate": 2.0600960135216462e-05,
"loss": 0.3446,
"step": 792
},
{
"epoch": 0.0849127315558411,
"grad_norm": 0.1462964117527008,
"learning_rate": 2.0410365331898416e-05,
"loss": 0.2856,
"step": 793
},
{
"epoch": 0.08501980940143485,
"grad_norm": 0.1861981898546219,
"learning_rate": 2.0220556046142893e-05,
"loss": 0.3894,
"step": 794
},
{
"epoch": 0.08512688724702859,
"grad_norm": 0.15652887523174286,
"learning_rate": 2.0031534151290943e-05,
"loss": 0.245,
"step": 795
},
{
"epoch": 0.08523396509262234,
"grad_norm": 0.14712314307689667,
"learning_rate": 1.9843301512912327e-05,
"loss": 0.3458,
"step": 796
},
{
"epoch": 0.08534104293821608,
"grad_norm": 0.15813574194908142,
"learning_rate": 1.965585998878724e-05,
"loss": 0.3723,
"step": 797
},
{
"epoch": 0.08544812078380983,
"grad_norm": 0.1604684740304947,
"learning_rate": 1.946921142888781e-05,
"loss": 0.3116,
"step": 798
},
{
"epoch": 0.08555519862940357,
"grad_norm": 0.14358116686344147,
"learning_rate": 1.928335767535997e-05,
"loss": 0.2253,
"step": 799
},
{
"epoch": 0.08566227647499733,
"grad_norm": 0.17999866604804993,
"learning_rate": 1.9098300562505266e-05,
"loss": 0.3825,
"step": 800
},
{
"epoch": 0.08576935432059107,
"grad_norm": 0.1776675432920456,
"learning_rate": 1.891404191676265e-05,
"loss": 0.3816,
"step": 801
},
{
"epoch": 0.08587643216618482,
"grad_norm": 0.16066551208496094,
"learning_rate": 1.8730583556690605e-05,
"loss": 0.2955,
"step": 802
},
{
"epoch": 0.08598351001177856,
"grad_norm": 0.1738562434911728,
"learning_rate": 1.854792729294905e-05,
"loss": 0.1947,
"step": 803
},
{
"epoch": 0.08609058785737231,
"grad_norm": 0.18131086230278015,
"learning_rate": 1.8366074928281607e-05,
"loss": 0.2306,
"step": 804
},
{
"epoch": 0.08619766570296605,
"grad_norm": 0.17308153212070465,
"learning_rate": 1.818502825749764e-05,
"loss": 0.3446,
"step": 805
},
{
"epoch": 0.0863047435485598,
"grad_norm": 0.1967420130968094,
"learning_rate": 1.8004789067454764e-05,
"loss": 0.3794,
"step": 806
},
{
"epoch": 0.08641182139415354,
"grad_norm": 0.17925085127353668,
"learning_rate": 1.7825359137040988e-05,
"loss": 0.3216,
"step": 807
},
{
"epoch": 0.0865188992397473,
"grad_norm": 0.19170551002025604,
"learning_rate": 1.7646740237157256e-05,
"loss": 0.3629,
"step": 808
},
{
"epoch": 0.08662597708534105,
"grad_norm": 0.16575735807418823,
"learning_rate": 1.7468934130700044e-05,
"loss": 0.204,
"step": 809
},
{
"epoch": 0.08673305493093479,
"grad_norm": 0.15411527454853058,
"learning_rate": 1.7291942572543807e-05,
"loss": 0.2983,
"step": 810
},
{
"epoch": 0.08684013277652854,
"grad_norm": 0.1592104583978653,
"learning_rate": 1.7115767309523812e-05,
"loss": 0.3134,
"step": 811
},
{
"epoch": 0.08694721062212228,
"grad_norm": 0.19162140786647797,
"learning_rate": 1.6940410080418723e-05,
"loss": 0.4155,
"step": 812
},
{
"epoch": 0.08705428846771603,
"grad_norm": 0.13956181704998016,
"learning_rate": 1.6765872615933677e-05,
"loss": 0.248,
"step": 813
},
{
"epoch": 0.08716136631330977,
"grad_norm": 0.22904083132743835,
"learning_rate": 1.6592156638682886e-05,
"loss": 0.3456,
"step": 814
},
{
"epoch": 0.08726844415890353,
"grad_norm": 0.15575331449508667,
"learning_rate": 1.6419263863172997e-05,
"loss": 0.2599,
"step": 815
},
{
"epoch": 0.08737552200449727,
"grad_norm": 0.17155703902244568,
"learning_rate": 1.6247195995785837e-05,
"loss": 0.3706,
"step": 816
},
{
"epoch": 0.08748259985009102,
"grad_norm": 0.1934448927640915,
"learning_rate": 1.6075954734761845e-05,
"loss": 0.3689,
"step": 817
},
{
"epoch": 0.08758967769568476,
"grad_norm": 0.14852175116539001,
"learning_rate": 1.5905541770183096e-05,
"loss": 0.257,
"step": 818
},
{
"epoch": 0.08769675554127851,
"grad_norm": 0.17280475795269012,
"learning_rate": 1.5735958783956794e-05,
"loss": 0.3368,
"step": 819
},
{
"epoch": 0.08780383338687225,
"grad_norm": 0.18640533089637756,
"learning_rate": 1.5567207449798515e-05,
"loss": 0.3615,
"step": 820
},
{
"epoch": 0.087910911232466,
"grad_norm": 0.19203537702560425,
"learning_rate": 1.539928943321579e-05,
"loss": 0.4128,
"step": 821
},
{
"epoch": 0.08801798907805974,
"grad_norm": 0.1681956797838211,
"learning_rate": 1.5232206391491699e-05,
"loss": 0.3007,
"step": 822
},
{
"epoch": 0.0881250669236535,
"grad_norm": 0.167117178440094,
"learning_rate": 1.5065959973668353e-05,
"loss": 0.3348,
"step": 823
},
{
"epoch": 0.08823214476924725,
"grad_norm": 0.145122230052948,
"learning_rate": 1.4900551820530828e-05,
"loss": 0.3176,
"step": 824
},
{
"epoch": 0.08833922261484099,
"grad_norm": 0.2096463143825531,
"learning_rate": 1.4735983564590783e-05,
"loss": 0.4623,
"step": 825
},
{
"epoch": 0.08844630046043474,
"grad_norm": 0.17497724294662476,
"learning_rate": 1.4572256830070497e-05,
"loss": 0.4211,
"step": 826
},
{
"epoch": 0.08855337830602848,
"grad_norm": 0.16783024370670319,
"learning_rate": 1.4409373232886702e-05,
"loss": 0.3312,
"step": 827
},
{
"epoch": 0.08866045615162223,
"grad_norm": 0.17440426349639893,
"learning_rate": 1.4247334380634792e-05,
"loss": 0.3737,
"step": 828
},
{
"epoch": 0.08876753399721597,
"grad_norm": 0.25179216265678406,
"learning_rate": 1.4086141872572789e-05,
"loss": 0.4219,
"step": 829
},
{
"epoch": 0.08887461184280972,
"grad_norm": 0.15539956092834473,
"learning_rate": 1.3925797299605647e-05,
"loss": 0.2873,
"step": 830
},
{
"epoch": 0.08898168968840348,
"grad_norm": 0.19273389875888824,
"learning_rate": 1.3766302244269624e-05,
"loss": 0.4413,
"step": 831
},
{
"epoch": 0.08908876753399722,
"grad_norm": 0.15699490904808044,
"learning_rate": 1.3607658280716473e-05,
"loss": 0.3176,
"step": 832
},
{
"epoch": 0.08919584537959097,
"grad_norm": 0.16093392670154572,
"learning_rate": 1.3449866974698122e-05,
"loss": 0.2945,
"step": 833
},
{
"epoch": 0.08930292322518471,
"grad_norm": 0.13421513140201569,
"learning_rate": 1.3292929883550998e-05,
"loss": 0.2489,
"step": 834
},
{
"epoch": 0.08941000107077846,
"grad_norm": 0.1671517938375473,
"learning_rate": 1.3136848556180892e-05,
"loss": 0.3242,
"step": 835
},
{
"epoch": 0.0895170789163722,
"grad_norm": 0.19687211513519287,
"learning_rate": 1.2981624533047432e-05,
"loss": 0.4313,
"step": 836
},
{
"epoch": 0.08962415676196595,
"grad_norm": 0.17762812972068787,
"learning_rate": 1.2827259346149122e-05,
"loss": 0.3867,
"step": 837
},
{
"epoch": 0.08973123460755969,
"grad_norm": 0.12915226817131042,
"learning_rate": 1.2673754519008008e-05,
"loss": 0.2702,
"step": 838
},
{
"epoch": 0.08983831245315345,
"grad_norm": 0.1644250750541687,
"learning_rate": 1.2521111566654731e-05,
"loss": 0.2511,
"step": 839
},
{
"epoch": 0.0899453902987472,
"grad_norm": 0.14260394871234894,
"learning_rate": 1.2369331995613665e-05,
"loss": 0.2988,
"step": 840
},
{
"epoch": 0.09005246814434094,
"grad_norm": 0.1875654011964798,
"learning_rate": 1.2218417303887842e-05,
"loss": 0.3031,
"step": 841
},
{
"epoch": 0.09015954598993468,
"grad_norm": 0.17389431595802307,
"learning_rate": 1.206836898094439e-05,
"loss": 0.322,
"step": 842
},
{
"epoch": 0.09026662383552843,
"grad_norm": 0.18567897379398346,
"learning_rate": 1.191918850769964e-05,
"loss": 0.3624,
"step": 843
},
{
"epoch": 0.09037370168112217,
"grad_norm": 0.23172374069690704,
"learning_rate": 1.1770877356504683e-05,
"loss": 0.4064,
"step": 844
},
{
"epoch": 0.09048077952671592,
"grad_norm": 0.17461536824703217,
"learning_rate": 1.1623436991130654e-05,
"loss": 0.3287,
"step": 845
},
{
"epoch": 0.09058785737230966,
"grad_norm": 0.15354901552200317,
"learning_rate": 1.1476868866754486e-05,
"loss": 0.2966,
"step": 846
},
{
"epoch": 0.09069493521790342,
"grad_norm": 0.13990119099617004,
"learning_rate": 1.1331174429944347e-05,
"loss": 0.2597,
"step": 847
},
{
"epoch": 0.09080201306349717,
"grad_norm": 0.15335947275161743,
"learning_rate": 1.1186355118645554e-05,
"loss": 0.3011,
"step": 848
},
{
"epoch": 0.09090909090909091,
"grad_norm": 0.164723739027977,
"learning_rate": 1.1042412362166222e-05,
"loss": 0.2423,
"step": 849
},
{
"epoch": 0.09101616875468466,
"grad_norm": 0.18542508780956268,
"learning_rate": 1.0899347581163221e-05,
"loss": 0.372,
"step": 850
},
{
"epoch": 0.0911232466002784,
"grad_norm": 0.16289161145687103,
"learning_rate": 1.0757162187628222e-05,
"loss": 0.3027,
"step": 851
},
{
"epoch": 0.09123032444587215,
"grad_norm": 0.17643985152244568,
"learning_rate": 1.0615857584873623e-05,
"loss": 0.3172,
"step": 852
},
{
"epoch": 0.09133740229146589,
"grad_norm": 0.1658363938331604,
"learning_rate": 1.0475435167518843e-05,
"loss": 0.2663,
"step": 853
},
{
"epoch": 0.09144448013705964,
"grad_norm": 0.17886339128017426,
"learning_rate": 1.0335896321476413e-05,
"loss": 0.4144,
"step": 854
},
{
"epoch": 0.0915515579826534,
"grad_norm": 0.15143504738807678,
"learning_rate": 1.0197242423938446e-05,
"loss": 0.2608,
"step": 855
},
{
"epoch": 0.09165863582824714,
"grad_norm": 0.2437451332807541,
"learning_rate": 1.0059474843362892e-05,
"loss": 0.3112,
"step": 856
},
{
"epoch": 0.09176571367384088,
"grad_norm": 0.2071426510810852,
"learning_rate": 9.922594939460194e-06,
"loss": 0.434,
"step": 857
},
{
"epoch": 0.09187279151943463,
"grad_norm": 0.15184633433818817,
"learning_rate": 9.786604063179728e-06,
"loss": 0.3421,
"step": 858
},
{
"epoch": 0.09197986936502837,
"grad_norm": 0.20159287750720978,
"learning_rate": 9.651503556696516e-06,
"loss": 0.3295,
"step": 859
},
{
"epoch": 0.09208694721062212,
"grad_norm": 0.1514168232679367,
"learning_rate": 9.517294753398064e-06,
"loss": 0.2855,
"step": 860
},
{
"epoch": 0.09219402505621586,
"grad_norm": 0.1669357568025589,
"learning_rate": 9.383978977871021e-06,
"loss": 0.254,
"step": 861
},
{
"epoch": 0.09230110290180961,
"grad_norm": 0.17768973112106323,
"learning_rate": 9.251557545888312e-06,
"loss": 0.4355,
"step": 862
},
{
"epoch": 0.09240818074740337,
"grad_norm": 0.17250321805477142,
"learning_rate": 9.120031764395987e-06,
"loss": 0.312,
"step": 863
},
{
"epoch": 0.09251525859299711,
"grad_norm": 0.17513404786586761,
"learning_rate": 8.989402931500434e-06,
"loss": 0.3356,
"step": 864
},
{
"epoch": 0.09262233643859086,
"grad_norm": 0.15594890713691711,
"learning_rate": 8.85967233645547e-06,
"loss": 0.2917,
"step": 865
},
{
"epoch": 0.0927294142841846,
"grad_norm": 0.14464133977890015,
"learning_rate": 8.730841259649725e-06,
"loss": 0.2637,
"step": 866
},
{
"epoch": 0.09283649212977835,
"grad_norm": 0.16818121075630188,
"learning_rate": 8.602910972593892e-06,
"loss": 0.3922,
"step": 867
},
{
"epoch": 0.09294356997537209,
"grad_norm": 0.16253922879695892,
"learning_rate": 8.475882737908248e-06,
"loss": 0.3469,
"step": 868
},
{
"epoch": 0.09305064782096584,
"grad_norm": 0.18424832820892334,
"learning_rate": 8.34975780931021e-06,
"loss": 0.3787,
"step": 869
},
{
"epoch": 0.09315772566655958,
"grad_norm": 0.1491045355796814,
"learning_rate": 8.224537431601886e-06,
"loss": 0.2689,
"step": 870
},
{
"epoch": 0.09326480351215334,
"grad_norm": 0.19525621831417084,
"learning_rate": 8.100222840657878e-06,
"loss": 0.3922,
"step": 871
},
{
"epoch": 0.09337188135774709,
"grad_norm": 0.20885202288627625,
"learning_rate": 7.976815263412963e-06,
"loss": 0.3052,
"step": 872
},
{
"epoch": 0.09347895920334083,
"grad_norm": 0.15215760469436646,
"learning_rate": 7.854315917850163e-06,
"loss": 0.2968,
"step": 873
},
{
"epoch": 0.09358603704893458,
"grad_norm": 0.17333446443080902,
"learning_rate": 7.73272601298851e-06,
"loss": 0.2817,
"step": 874
},
{
"epoch": 0.09369311489452832,
"grad_norm": 0.17127728462219238,
"learning_rate": 7.612046748871327e-06,
"loss": 0.2874,
"step": 875
},
{
"epoch": 0.09380019274012207,
"grad_norm": 0.16041713953018188,
"learning_rate": 7.492279316554207e-06,
"loss": 0.3136,
"step": 876
},
{
"epoch": 0.09390727058571581,
"grad_norm": 0.1982324719429016,
"learning_rate": 7.3734248980933395e-06,
"loss": 0.3975,
"step": 877
},
{
"epoch": 0.09401434843130957,
"grad_norm": 0.15895456075668335,
"learning_rate": 7.255484666533874e-06,
"loss": 0.3546,
"step": 878
},
{
"epoch": 0.09412142627690331,
"grad_norm": 0.19794389605522156,
"learning_rate": 7.138459785898266e-06,
"loss": 0.4034,
"step": 879
},
{
"epoch": 0.09422850412249706,
"grad_norm": 0.15700578689575195,
"learning_rate": 7.022351411174866e-06,
"loss": 0.3068,
"step": 880
},
{
"epoch": 0.0943355819680908,
"grad_norm": 0.17890901863574982,
"learning_rate": 6.907160688306425e-06,
"loss": 0.2851,
"step": 881
},
{
"epoch": 0.09444265981368455,
"grad_norm": 0.18140609562397003,
"learning_rate": 6.7928887541789055e-06,
"loss": 0.3212,
"step": 882
},
{
"epoch": 0.0945497376592783,
"grad_norm": 0.13674962520599365,
"learning_rate": 6.679536736610137e-06,
"loss": 0.2776,
"step": 883
},
{
"epoch": 0.09465681550487204,
"grad_norm": 0.13583561778068542,
"learning_rate": 6.5671057543387985e-06,
"loss": 0.2168,
"step": 884
},
{
"epoch": 0.09476389335046578,
"grad_norm": 0.19280678033828735,
"learning_rate": 6.455596917013273e-06,
"loss": 0.2604,
"step": 885
},
{
"epoch": 0.09487097119605954,
"grad_norm": 0.16224633157253265,
"learning_rate": 6.345011325180772e-06,
"loss": 0.3116,
"step": 886
},
{
"epoch": 0.09497804904165329,
"grad_norm": 0.1368846297264099,
"learning_rate": 6.235350070276447e-06,
"loss": 0.252,
"step": 887
},
{
"epoch": 0.09508512688724703,
"grad_norm": 0.12165465205907822,
"learning_rate": 6.126614234612593e-06,
"loss": 0.2395,
"step": 888
},
{
"epoch": 0.09519220473284078,
"grad_norm": 0.1612234264612198,
"learning_rate": 6.018804891368035e-06,
"loss": 0.3441,
"step": 889
},
{
"epoch": 0.09529928257843452,
"grad_norm": 0.15797550976276398,
"learning_rate": 5.911923104577455e-06,
"loss": 0.3158,
"step": 890
},
{
"epoch": 0.09540636042402827,
"grad_norm": 0.17879457771778107,
"learning_rate": 5.805969929120947e-06,
"loss": 0.3497,
"step": 891
},
{
"epoch": 0.09551343826962201,
"grad_norm": 0.1826862245798111,
"learning_rate": 5.700946410713548e-06,
"loss": 0.4023,
"step": 892
},
{
"epoch": 0.09562051611521576,
"grad_norm": 0.14438298344612122,
"learning_rate": 5.5968535858950345e-06,
"loss": 0.2356,
"step": 893
},
{
"epoch": 0.09572759396080951,
"grad_norm": 0.1626652181148529,
"learning_rate": 5.49369248201953e-06,
"loss": 0.2726,
"step": 894
},
{
"epoch": 0.09583467180640326,
"grad_norm": 0.235837921500206,
"learning_rate": 5.39146411724547e-06,
"loss": 0.4375,
"step": 895
},
{
"epoch": 0.095941749651997,
"grad_norm": 0.16252082586288452,
"learning_rate": 5.290169500525577e-06,
"loss": 0.3127,
"step": 896
},
{
"epoch": 0.09604882749759075,
"grad_norm": 0.20119427144527435,
"learning_rate": 5.189809631596798e-06,
"loss": 0.4466,
"step": 897
},
{
"epoch": 0.0961559053431845,
"grad_norm": 0.15925422310829163,
"learning_rate": 5.0903855009705514e-06,
"loss": 0.2865,
"step": 898
},
{
"epoch": 0.09626298318877824,
"grad_norm": 0.15317882597446442,
"learning_rate": 4.991898089922819e-06,
"loss": 0.336,
"step": 899
},
{
"epoch": 0.09637006103437198,
"grad_norm": 0.1327366828918457,
"learning_rate": 4.8943483704846475e-06,
"loss": 0.2251,
"step": 900
},
{
"epoch": 0.09647713887996573,
"grad_norm": 0.16071724891662598,
"learning_rate": 4.797737305432337e-06,
"loss": 0.307,
"step": 901
},
{
"epoch": 0.09658421672555949,
"grad_norm": 0.16794702410697937,
"learning_rate": 4.702065848278126e-06,
"loss": 0.3508,
"step": 902
},
{
"epoch": 0.09669129457115323,
"grad_norm": 0.13818074762821198,
"learning_rate": 4.607334943260655e-06,
"loss": 0.2423,
"step": 903
},
{
"epoch": 0.09679837241674698,
"grad_norm": 0.20473352074623108,
"learning_rate": 4.513545525335705e-06,
"loss": 0.4169,
"step": 904
},
{
"epoch": 0.09690545026234072,
"grad_norm": 0.1584278792142868,
"learning_rate": 4.420698520166988e-06,
"loss": 0.3027,
"step": 905
},
{
"epoch": 0.09701252810793447,
"grad_norm": 0.1746038943529129,
"learning_rate": 4.328794844116946e-06,
"loss": 0.2906,
"step": 906
},
{
"epoch": 0.09711960595352821,
"grad_norm": 0.17455877363681793,
"learning_rate": 4.237835404237778e-06,
"loss": 0.2776,
"step": 907
},
{
"epoch": 0.09722668379912196,
"grad_norm": 0.1757044494152069,
"learning_rate": 4.147821098262405e-06,
"loss": 0.3792,
"step": 908
},
{
"epoch": 0.0973337616447157,
"grad_norm": 0.1343322992324829,
"learning_rate": 4.0587528145957235e-06,
"loss": 0.2821,
"step": 909
},
{
"epoch": 0.09744083949030946,
"grad_norm": 0.17307552695274353,
"learning_rate": 3.970631432305694e-06,
"loss": 0.3373,
"step": 910
},
{
"epoch": 0.0975479173359032,
"grad_norm": 0.16109535098075867,
"learning_rate": 3.883457821114811e-06,
"loss": 0.2397,
"step": 911
},
{
"epoch": 0.09765499518149695,
"grad_norm": 0.17878471314907074,
"learning_rate": 3.797232841391407e-06,
"loss": 0.4117,
"step": 912
},
{
"epoch": 0.0977620730270907,
"grad_norm": 0.1923290491104126,
"learning_rate": 3.711957344141237e-06,
"loss": 0.3319,
"step": 913
},
{
"epoch": 0.09786915087268444,
"grad_norm": 0.2836756110191345,
"learning_rate": 3.627632170999029e-06,
"loss": 0.2799,
"step": 914
},
{
"epoch": 0.09797622871827819,
"grad_norm": 0.171551913022995,
"learning_rate": 3.5442581542201923e-06,
"loss": 0.3929,
"step": 915
},
{
"epoch": 0.09808330656387193,
"grad_norm": 0.15683524310588837,
"learning_rate": 3.461836116672612e-06,
"loss": 0.28,
"step": 916
},
{
"epoch": 0.09819038440946568,
"grad_norm": 0.17995594441890717,
"learning_rate": 3.380366871828522e-06,
"loss": 0.4169,
"step": 917
},
{
"epoch": 0.09829746225505943,
"grad_norm": 0.16825750470161438,
"learning_rate": 3.2998512237565005e-06,
"loss": 0.2996,
"step": 918
},
{
"epoch": 0.09840454010065318,
"grad_norm": 0.1549074649810791,
"learning_rate": 3.2202899671134546e-06,
"loss": 0.3138,
"step": 919
},
{
"epoch": 0.09851161794624692,
"grad_norm": 0.15720857679843903,
"learning_rate": 3.1416838871368924e-06,
"loss": 0.2918,
"step": 920
},
{
"epoch": 0.09861869579184067,
"grad_norm": 0.16377247869968414,
"learning_rate": 3.064033759637064e-06,
"loss": 0.3749,
"step": 921
},
{
"epoch": 0.09872577363743441,
"grad_norm": 0.18883401155471802,
"learning_rate": 2.9873403509894203e-06,
"loss": 0.3754,
"step": 922
},
{
"epoch": 0.09883285148302816,
"grad_norm": 0.19733136892318726,
"learning_rate": 2.9116044181269007e-06,
"loss": 0.3307,
"step": 923
},
{
"epoch": 0.0989399293286219,
"grad_norm": 0.20441798865795135,
"learning_rate": 2.836826708532603e-06,
"loss": 0.3909,
"step": 924
},
{
"epoch": 0.09904700717421565,
"grad_norm": 0.1249212846159935,
"learning_rate": 2.7630079602323442e-06,
"loss": 0.2121,
"step": 925
},
{
"epoch": 0.09915408501980941,
"grad_norm": 0.2168767899274826,
"learning_rate": 2.690148901787337e-06,
"loss": 0.3815,
"step": 926
},
{
"epoch": 0.09926116286540315,
"grad_norm": 0.15949676930904388,
"learning_rate": 2.618250252287113e-06,
"loss": 0.3354,
"step": 927
},
{
"epoch": 0.0993682407109969,
"grad_norm": 0.20438000559806824,
"learning_rate": 2.5473127213422763e-06,
"loss": 0.3388,
"step": 928
},
{
"epoch": 0.09947531855659064,
"grad_norm": 0.17350825667381287,
"learning_rate": 2.4773370090776626e-06,
"loss": 0.2954,
"step": 929
},
{
"epoch": 0.09958239640218439,
"grad_norm": 0.19200490415096283,
"learning_rate": 2.4083238061252567e-06,
"loss": 0.5093,
"step": 930
},
{
"epoch": 0.09968947424777813,
"grad_norm": 0.1551746129989624,
"learning_rate": 2.3402737936175425e-06,
"loss": 0.3675,
"step": 931
},
{
"epoch": 0.09979655209337188,
"grad_norm": 0.16438502073287964,
"learning_rate": 2.273187643180652e-06,
"loss": 0.3649,
"step": 932
},
{
"epoch": 0.09990362993896562,
"grad_norm": 0.2188139408826828,
"learning_rate": 2.2070660169278166e-06,
"loss": 0.3949,
"step": 933
},
{
"epoch": 0.10001070778455938,
"grad_norm": 0.14059962332248688,
"learning_rate": 2.141909567452793e-06,
"loss": 0.256,
"step": 934
},
{
"epoch": 0.10011778563015312,
"grad_norm": 0.15545494854450226,
"learning_rate": 2.0777189378234143e-06,
"loss": 0.3074,
"step": 935
},
{
"epoch": 0.10022486347574687,
"grad_norm": 0.1423916071653366,
"learning_rate": 2.014494761575314e-06,
"loss": 0.2434,
"step": 936
},
{
"epoch": 0.10033194132134061,
"grad_norm": 0.14946754276752472,
"learning_rate": 1.9522376627055583e-06,
"loss": 0.3399,
"step": 937
},
{
"epoch": 0.10043901916693436,
"grad_norm": 0.1648038625717163,
"learning_rate": 1.8909482556666024e-06,
"loss": 0.3306,
"step": 938
},
{
"epoch": 0.1005460970125281,
"grad_norm": 0.17612142860889435,
"learning_rate": 1.8306271453601199e-06,
"loss": 0.3385,
"step": 939
},
{
"epoch": 0.10065317485812185,
"grad_norm": 0.13343556225299835,
"learning_rate": 1.771274927131139e-06,
"loss": 0.2621,
"step": 940
},
{
"epoch": 0.10076025270371561,
"grad_norm": 0.1661439836025238,
"learning_rate": 1.712892186762083e-06,
"loss": 0.3304,
"step": 941
},
{
"epoch": 0.10086733054930935,
"grad_norm": 0.1674559861421585,
"learning_rate": 1.6554795004670388e-06,
"loss": 0.4132,
"step": 942
},
{
"epoch": 0.1009744083949031,
"grad_norm": 0.2064605951309204,
"learning_rate": 1.5990374348860305e-06,
"loss": 0.3322,
"step": 943
},
{
"epoch": 0.10108148624049684,
"grad_norm": 0.1431376039981842,
"learning_rate": 1.543566547079467e-06,
"loss": 0.2727,
"step": 944
},
{
"epoch": 0.10118856408609059,
"grad_norm": 0.19648565351963043,
"learning_rate": 1.4890673845226133e-06,
"loss": 0.4002,
"step": 945
},
{
"epoch": 0.10129564193168433,
"grad_norm": 0.1771528124809265,
"learning_rate": 1.4355404851001952e-06,
"loss": 0.4174,
"step": 946
},
{
"epoch": 0.10140271977727808,
"grad_norm": 0.16234669089317322,
"learning_rate": 1.3829863771011253e-06,
"loss": 0.2401,
"step": 947
},
{
"epoch": 0.10150979762287182,
"grad_norm": 0.17918367683887482,
"learning_rate": 1.3314055792131964e-06,
"loss": 0.3435,
"step": 948
},
{
"epoch": 0.10161687546846558,
"grad_norm": 0.17307892441749573,
"learning_rate": 1.280798600518085e-06,
"loss": 0.3192,
"step": 949
},
{
"epoch": 0.10172395331405933,
"grad_norm": 0.15263895690441132,
"learning_rate": 1.231165940486234e-06,
"loss": 0.2911,
"step": 950
},
{
"epoch": 0.10183103115965307,
"grad_norm": 0.1903427690267563,
"learning_rate": 1.1825080889719563e-06,
"loss": 0.3734,
"step": 951
},
{
"epoch": 0.10193810900524682,
"grad_norm": 0.14863133430480957,
"learning_rate": 1.134825526208605e-06,
"loss": 0.2192,
"step": 952
},
{
"epoch": 0.10204518685084056,
"grad_norm": 0.14327697455883026,
"learning_rate": 1.0881187228038215e-06,
"loss": 0.2801,
"step": 953
},
{
"epoch": 0.1021522646964343,
"grad_norm": 0.16685131192207336,
"learning_rate": 1.0423881397349068e-06,
"loss": 0.3592,
"step": 954
},
{
"epoch": 0.10225934254202805,
"grad_norm": 0.16818654537200928,
"learning_rate": 9.976342283442463e-07,
"loss": 0.323,
"step": 955
},
{
"epoch": 0.1023664203876218,
"grad_norm": 0.20931144058704376,
"learning_rate": 9.538574303348813e-07,
"loss": 0.3971,
"step": 956
},
{
"epoch": 0.10247349823321555,
"grad_norm": 0.1491042673587799,
"learning_rate": 9.110581777661331e-07,
"loss": 0.2972,
"step": 957
},
{
"epoch": 0.1025805760788093,
"grad_norm": 0.14343906939029694,
"learning_rate": 8.692368930493521e-07,
"loss": 0.261,
"step": 958
},
{
"epoch": 0.10268765392440304,
"grad_norm": 0.1593460738658905,
"learning_rate": 8.283939889437209e-07,
"loss": 0.3349,
"step": 959
},
{
"epoch": 0.10279473176999679,
"grad_norm": 0.16361872851848602,
"learning_rate": 7.885298685522235e-07,
"loss": 0.3133,
"step": 960
},
{
"epoch": 0.10290180961559053,
"grad_norm": 0.20151600241661072,
"learning_rate": 7.496449253176274e-07,
"loss": 0.3668,
"step": 961
},
{
"epoch": 0.10300888746118428,
"grad_norm": 0.17242741584777832,
"learning_rate": 7.117395430186414e-07,
"loss": 0.3549,
"step": 962
},
{
"epoch": 0.10311596530677802,
"grad_norm": 0.21072721481323242,
"learning_rate": 6.748140957660631e-07,
"loss": 0.3799,
"step": 963
},
{
"epoch": 0.10322304315237177,
"grad_norm": 0.18324029445648193,
"learning_rate": 6.388689479991605e-07,
"loss": 0.3466,
"step": 964
},
{
"epoch": 0.10333012099796553,
"grad_norm": 0.13959814608097076,
"learning_rate": 6.039044544820404e-07,
"loss": 0.2795,
"step": 965
},
{
"epoch": 0.10343719884355927,
"grad_norm": 0.13971494138240814,
"learning_rate": 5.699209603001076e-07,
"loss": 0.3022,
"step": 966
},
{
"epoch": 0.10354427668915302,
"grad_norm": 0.1750052124261856,
"learning_rate": 5.369188008567672e-07,
"loss": 0.3426,
"step": 967
},
{
"epoch": 0.10365135453474676,
"grad_norm": 0.17160281538963318,
"learning_rate": 5.048983018699827e-07,
"loss": 0.3439,
"step": 968
},
{
"epoch": 0.1037584323803405,
"grad_norm": 0.17078298330307007,
"learning_rate": 4.738597793691679e-07,
"loss": 0.34,
"step": 969
},
{
"epoch": 0.10386551022593425,
"grad_norm": 0.1592579185962677,
"learning_rate": 4.438035396920004e-07,
"loss": 0.2662,
"step": 970
},
{
"epoch": 0.103972588071528,
"grad_norm": 0.23051397502422333,
"learning_rate": 4.1472987948143473e-07,
"loss": 0.5427,
"step": 971
},
{
"epoch": 0.10407966591712174,
"grad_norm": 0.14807277917861938,
"learning_rate": 3.866390856827495e-07,
"loss": 0.3386,
"step": 972
},
{
"epoch": 0.1041867437627155,
"grad_norm": 0.17786002159118652,
"learning_rate": 3.595314355407609e-07,
"loss": 0.3552,
"step": 973
},
{
"epoch": 0.10429382160830925,
"grad_norm": 0.14650464057922363,
"learning_rate": 3.3340719659701313e-07,
"loss": 0.309,
"step": 974
},
{
"epoch": 0.10440089945390299,
"grad_norm": 0.16224049031734467,
"learning_rate": 3.0826662668720364e-07,
"loss": 0.3336,
"step": 975
},
{
"epoch": 0.10450797729949673,
"grad_norm": 0.1791958510875702,
"learning_rate": 2.841099739386066e-07,
"loss": 0.3769,
"step": 976
},
{
"epoch": 0.10461505514509048,
"grad_norm": 0.154682457447052,
"learning_rate": 2.609374767676309e-07,
"loss": 0.3388,
"step": 977
},
{
"epoch": 0.10472213299068422,
"grad_norm": 0.2175721973180771,
"learning_rate": 2.387493638774774e-07,
"loss": 0.4477,
"step": 978
},
{
"epoch": 0.10482921083627797,
"grad_norm": 0.16635477542877197,
"learning_rate": 2.175458542558517e-07,
"loss": 0.3834,
"step": 979
},
{
"epoch": 0.10493628868187171,
"grad_norm": 0.17459362745285034,
"learning_rate": 1.973271571728441e-07,
"loss": 0.288,
"step": 980
},
{
"epoch": 0.10504336652746547,
"grad_norm": 0.16596029698848724,
"learning_rate": 1.7809347217881966e-07,
"loss": 0.3816,
"step": 981
},
{
"epoch": 0.10515044437305922,
"grad_norm": 0.18052440881729126,
"learning_rate": 1.598449891024978e-07,
"loss": 0.3266,
"step": 982
},
{
"epoch": 0.10525752221865296,
"grad_norm": 0.17302225530147552,
"learning_rate": 1.425818880490315e-07,
"loss": 0.3603,
"step": 983
},
{
"epoch": 0.10536460006424671,
"grad_norm": 0.20757171511650085,
"learning_rate": 1.2630433939825327e-07,
"loss": 0.3392,
"step": 984
},
{
"epoch": 0.10547167790984045,
"grad_norm": 0.15444600582122803,
"learning_rate": 1.1101250380300965e-07,
"loss": 0.3041,
"step": 985
},
{
"epoch": 0.1055787557554342,
"grad_norm": 0.17258261144161224,
"learning_rate": 9.670653218752934e-08,
"loss": 0.3568,
"step": 986
},
{
"epoch": 0.10568583360102794,
"grad_norm": 0.1691325157880783,
"learning_rate": 8.33865657459909e-08,
"loss": 0.4094,
"step": 987
},
{
"epoch": 0.10579291144662169,
"grad_norm": 0.204833522439003,
"learning_rate": 7.105273594107953e-08,
"loss": 0.3808,
"step": 988
},
{
"epoch": 0.10589998929221545,
"grad_norm": 0.1730750948190689,
"learning_rate": 5.970516450271025e-08,
"loss": 0.3734,
"step": 989
},
{
"epoch": 0.10600706713780919,
"grad_norm": 0.16625729203224182,
"learning_rate": 4.934396342684e-08,
"loss": 0.3676,
"step": 990
},
{
"epoch": 0.10611414498340294,
"grad_norm": 0.15599259734153748,
"learning_rate": 3.996923497434635e-08,
"loss": 0.3668,
"step": 991
},
{
"epoch": 0.10622122282899668,
"grad_norm": 0.1792498528957367,
"learning_rate": 3.1581071670006015e-08,
"loss": 0.4077,
"step": 992
},
{
"epoch": 0.10632830067459043,
"grad_norm": 0.18159109354019165,
"learning_rate": 2.417955630159563e-08,
"loss": 0.3609,
"step": 993
},
{
"epoch": 0.10643537852018417,
"grad_norm": 0.1742103546857834,
"learning_rate": 1.7764761919103477e-08,
"loss": 0.3711,
"step": 994
},
{
"epoch": 0.10654245636577792,
"grad_norm": 0.15315918624401093,
"learning_rate": 1.2336751833941229e-08,
"loss": 0.2493,
"step": 995
},
{
"epoch": 0.10664953421137166,
"grad_norm": 0.1537967324256897,
"learning_rate": 7.895579618388827e-09,
"loss": 0.3098,
"step": 996
},
{
"epoch": 0.10675661205696542,
"grad_norm": 0.17367656528949738,
"learning_rate": 4.4412891050171765e-09,
"loss": 0.3143,
"step": 997
},
{
"epoch": 0.10686368990255916,
"grad_norm": 0.14059410989284515,
"learning_rate": 1.973914386288467e-09,
"loss": 0.2818,
"step": 998
},
{
"epoch": 0.10697076774815291,
"grad_norm": 0.21730297803878784,
"learning_rate": 4.934798141786879e-10,
"loss": 0.4292,
"step": 999
},
{
"epoch": 0.10707784559374665,
"grad_norm": 0.21436914801597595,
"learning_rate": 0.0,
"loss": 0.4271,
"step": 1000
},
{
"epoch": 0.10707784559374665,
"eval_loss": 0.33198028802871704,
"eval_runtime": 367.0984,
"eval_samples_per_second": 5.358,
"eval_steps_per_second": 5.358,
"step": 1000
}
],
"logging_steps": 1,
"max_steps": 1000,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.6666433267734938e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}