qwen / trainer_state.json
lokijii's picture
Upload 8 files
4a4ee49 verified
raw
history blame
189 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.9995792131285504,
"eval_steps": 500,
"global_step": 1188,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.001683147485798443,
"grad_norm": 0.17560942471027374,
"learning_rate": 0.0,
"loss": 2.613,
"step": 1
},
{
"epoch": 0.003366294971596886,
"grad_norm": 0.15861666202545166,
"learning_rate": 2.7894294565112984e-06,
"loss": 2.6655,
"step": 2
},
{
"epoch": 0.005049442457395329,
"grad_norm": 0.1817302405834198,
"learning_rate": 4.421141086977404e-06,
"loss": 2.55,
"step": 3
},
{
"epoch": 0.006732589943193772,
"grad_norm": 0.17854492366313934,
"learning_rate": 5.578858913022597e-06,
"loss": 2.7908,
"step": 4
},
{
"epoch": 0.008415737428992216,
"grad_norm": 0.17169038951396942,
"learning_rate": 6.47685462377997e-06,
"loss": 2.6868,
"step": 5
},
{
"epoch": 0.010098884914790659,
"grad_norm": 0.18368647992610931,
"learning_rate": 7.210570543488702e-06,
"loss": 2.5874,
"step": 6
},
{
"epoch": 0.011782032400589101,
"grad_norm": 0.19648714363574982,
"learning_rate": 7.830918514469461e-06,
"loss": 2.6633,
"step": 7
},
{
"epoch": 0.013465179886387544,
"grad_norm": 0.18358571827411652,
"learning_rate": 8.368288369533896e-06,
"loss": 2.6355,
"step": 8
},
{
"epoch": 0.015148327372185988,
"grad_norm": 0.19153611361980438,
"learning_rate": 8.842282173954808e-06,
"loss": 2.6633,
"step": 9
},
{
"epoch": 0.016831474857984433,
"grad_norm": 0.20646820962429047,
"learning_rate": 9.26628408029127e-06,
"loss": 2.7268,
"step": 10
},
{
"epoch": 0.018514622343782875,
"grad_norm": 0.18688935041427612,
"learning_rate": 9.64984045981344e-06,
"loss": 2.7832,
"step": 11
},
{
"epoch": 0.020197769829581318,
"grad_norm": 0.1985747218132019,
"learning_rate": 1e-05,
"loss": 2.738,
"step": 12
},
{
"epoch": 0.02188091731537976,
"grad_norm": 0.19321100413799286,
"learning_rate": 1e-05,
"loss": 2.6206,
"step": 13
},
{
"epoch": 0.023564064801178203,
"grad_norm": 0.1875382661819458,
"learning_rate": 1e-05,
"loss": 2.7153,
"step": 14
},
{
"epoch": 0.025247212286976645,
"grad_norm": 0.18803201615810394,
"learning_rate": 1e-05,
"loss": 2.5359,
"step": 15
},
{
"epoch": 0.026930359772775088,
"grad_norm": 0.19693922996520996,
"learning_rate": 1e-05,
"loss": 2.6082,
"step": 16
},
{
"epoch": 0.028613507258573534,
"grad_norm": 0.20534300804138184,
"learning_rate": 1e-05,
"loss": 2.5317,
"step": 17
},
{
"epoch": 0.030296654744371977,
"grad_norm": 0.22174465656280518,
"learning_rate": 1e-05,
"loss": 2.6067,
"step": 18
},
{
"epoch": 0.03197980223017042,
"grad_norm": 0.1947612464427948,
"learning_rate": 1e-05,
"loss": 2.6824,
"step": 19
},
{
"epoch": 0.033662949715968865,
"grad_norm": 0.19715926051139832,
"learning_rate": 1e-05,
"loss": 2.6868,
"step": 20
},
{
"epoch": 0.035346097201767304,
"grad_norm": 0.19586338102817535,
"learning_rate": 1e-05,
"loss": 2.6206,
"step": 21
},
{
"epoch": 0.03702924468756575,
"grad_norm": 0.19280074536800385,
"learning_rate": 1e-05,
"loss": 2.6023,
"step": 22
},
{
"epoch": 0.03871239217336419,
"grad_norm": 0.19658198952674866,
"learning_rate": 1e-05,
"loss": 2.6384,
"step": 23
},
{
"epoch": 0.040395539659162635,
"grad_norm": 0.17433768510818481,
"learning_rate": 1e-05,
"loss": 2.5305,
"step": 24
},
{
"epoch": 0.042078687144961074,
"grad_norm": 0.18013380467891693,
"learning_rate": 1e-05,
"loss": 2.6519,
"step": 25
},
{
"epoch": 0.04376183463075952,
"grad_norm": 0.1933555006980896,
"learning_rate": 1e-05,
"loss": 2.5591,
"step": 26
},
{
"epoch": 0.045444982116557966,
"grad_norm": 0.18386027216911316,
"learning_rate": 1e-05,
"loss": 2.6169,
"step": 27
},
{
"epoch": 0.047128129602356406,
"grad_norm": 0.18173415958881378,
"learning_rate": 1e-05,
"loss": 2.623,
"step": 28
},
{
"epoch": 0.04881127708815485,
"grad_norm": 0.19154761731624603,
"learning_rate": 1e-05,
"loss": 2.5981,
"step": 29
},
{
"epoch": 0.05049442457395329,
"grad_norm": 0.2001664638519287,
"learning_rate": 1e-05,
"loss": 2.5066,
"step": 30
},
{
"epoch": 0.05217757205975174,
"grad_norm": 0.15573543310165405,
"learning_rate": 1e-05,
"loss": 2.6013,
"step": 31
},
{
"epoch": 0.053860719545550176,
"grad_norm": 0.16071979701519012,
"learning_rate": 1e-05,
"loss": 2.4634,
"step": 32
},
{
"epoch": 0.05554386703134862,
"grad_norm": 0.1769736260175705,
"learning_rate": 1e-05,
"loss": 2.5491,
"step": 33
},
{
"epoch": 0.05722701451714707,
"grad_norm": 0.17623937129974365,
"learning_rate": 1e-05,
"loss": 2.4399,
"step": 34
},
{
"epoch": 0.05891016200294551,
"grad_norm": 0.17367449402809143,
"learning_rate": 1e-05,
"loss": 2.5464,
"step": 35
},
{
"epoch": 0.06059330948874395,
"grad_norm": 0.14842955768108368,
"learning_rate": 1e-05,
"loss": 2.4174,
"step": 36
},
{
"epoch": 0.06227645697454239,
"grad_norm": 0.17405100166797638,
"learning_rate": 1e-05,
"loss": 2.5303,
"step": 37
},
{
"epoch": 0.06395960446034084,
"grad_norm": 0.145203098654747,
"learning_rate": 1e-05,
"loss": 2.6428,
"step": 38
},
{
"epoch": 0.06564275194613928,
"grad_norm": 0.1542726755142212,
"learning_rate": 1e-05,
"loss": 2.5618,
"step": 39
},
{
"epoch": 0.06732589943193773,
"grad_norm": 0.14489781856536865,
"learning_rate": 1e-05,
"loss": 2.6885,
"step": 40
},
{
"epoch": 0.06900904691773617,
"grad_norm": 0.14798486232757568,
"learning_rate": 1e-05,
"loss": 2.5322,
"step": 41
},
{
"epoch": 0.07069219440353461,
"grad_norm": 0.15226829051971436,
"learning_rate": 1e-05,
"loss": 2.6011,
"step": 42
},
{
"epoch": 0.07237534188933305,
"grad_norm": 0.14561522006988525,
"learning_rate": 1e-05,
"loss": 2.5657,
"step": 43
},
{
"epoch": 0.0740584893751315,
"grad_norm": 0.13787826895713806,
"learning_rate": 1e-05,
"loss": 2.6011,
"step": 44
},
{
"epoch": 0.07574163686092994,
"grad_norm": 0.14005698263645172,
"learning_rate": 1e-05,
"loss": 2.4673,
"step": 45
},
{
"epoch": 0.07742478434672838,
"grad_norm": 0.13822345435619354,
"learning_rate": 1e-05,
"loss": 2.512,
"step": 46
},
{
"epoch": 0.07910793183252683,
"grad_norm": 0.1284177154302597,
"learning_rate": 1e-05,
"loss": 2.5625,
"step": 47
},
{
"epoch": 0.08079107931832527,
"grad_norm": 0.1279960423707962,
"learning_rate": 1e-05,
"loss": 2.46,
"step": 48
},
{
"epoch": 0.08247422680412371,
"grad_norm": 0.12479826807975769,
"learning_rate": 1e-05,
"loss": 2.5706,
"step": 49
},
{
"epoch": 0.08415737428992215,
"grad_norm": 0.12982836365699768,
"learning_rate": 1e-05,
"loss": 2.5098,
"step": 50
},
{
"epoch": 0.0858405217757206,
"grad_norm": 0.13269256055355072,
"learning_rate": 1e-05,
"loss": 2.4688,
"step": 51
},
{
"epoch": 0.08752366926151904,
"grad_norm": 0.11713477969169617,
"learning_rate": 1e-05,
"loss": 2.6226,
"step": 52
},
{
"epoch": 0.08920681674731748,
"grad_norm": 0.11179152131080627,
"learning_rate": 1e-05,
"loss": 2.4224,
"step": 53
},
{
"epoch": 0.09088996423311593,
"grad_norm": 0.12146276980638504,
"learning_rate": 1e-05,
"loss": 2.4639,
"step": 54
},
{
"epoch": 0.09257311171891437,
"grad_norm": 0.12470445781946182,
"learning_rate": 1e-05,
"loss": 2.5195,
"step": 55
},
{
"epoch": 0.09425625920471281,
"grad_norm": 0.11872275173664093,
"learning_rate": 1e-05,
"loss": 2.5186,
"step": 56
},
{
"epoch": 0.09593940669051125,
"grad_norm": 0.11616484075784683,
"learning_rate": 1e-05,
"loss": 2.5581,
"step": 57
},
{
"epoch": 0.0976225541763097,
"grad_norm": 0.1075875386595726,
"learning_rate": 1e-05,
"loss": 2.5693,
"step": 58
},
{
"epoch": 0.09930570166210814,
"grad_norm": 0.10176095366477966,
"learning_rate": 1e-05,
"loss": 2.521,
"step": 59
},
{
"epoch": 0.10098884914790658,
"grad_norm": 0.1076890155673027,
"learning_rate": 1e-05,
"loss": 2.53,
"step": 60
},
{
"epoch": 0.10267199663370503,
"grad_norm": 0.09105601906776428,
"learning_rate": 1e-05,
"loss": 2.3733,
"step": 61
},
{
"epoch": 0.10435514411950347,
"grad_norm": 0.09733142703771591,
"learning_rate": 1e-05,
"loss": 2.416,
"step": 62
},
{
"epoch": 0.10603829160530191,
"grad_norm": 0.09099874645471573,
"learning_rate": 1e-05,
"loss": 2.3774,
"step": 63
},
{
"epoch": 0.10772143909110035,
"grad_norm": 0.0884426161646843,
"learning_rate": 1e-05,
"loss": 2.4136,
"step": 64
},
{
"epoch": 0.1094045865768988,
"grad_norm": 0.08939989656209946,
"learning_rate": 1e-05,
"loss": 2.4482,
"step": 65
},
{
"epoch": 0.11108773406269724,
"grad_norm": 0.09078355878591537,
"learning_rate": 1e-05,
"loss": 2.5256,
"step": 66
},
{
"epoch": 0.11277088154849568,
"grad_norm": 0.08570227026939392,
"learning_rate": 1e-05,
"loss": 2.4954,
"step": 67
},
{
"epoch": 0.11445402903429414,
"grad_norm": 0.0766797736287117,
"learning_rate": 1e-05,
"loss": 2.3694,
"step": 68
},
{
"epoch": 0.11613717652009257,
"grad_norm": 0.08015618473291397,
"learning_rate": 1e-05,
"loss": 2.4724,
"step": 69
},
{
"epoch": 0.11782032400589101,
"grad_norm": 0.08956343680620193,
"learning_rate": 1e-05,
"loss": 2.47,
"step": 70
},
{
"epoch": 0.11950347149168945,
"grad_norm": 0.08134786039590836,
"learning_rate": 1e-05,
"loss": 2.4482,
"step": 71
},
{
"epoch": 0.1211866189774879,
"grad_norm": 0.07923366874456406,
"learning_rate": 1e-05,
"loss": 2.4182,
"step": 72
},
{
"epoch": 0.12286976646328635,
"grad_norm": 0.07909434288740158,
"learning_rate": 1e-05,
"loss": 2.3711,
"step": 73
},
{
"epoch": 0.12455291394908478,
"grad_norm": 0.07540368288755417,
"learning_rate": 1e-05,
"loss": 2.3962,
"step": 74
},
{
"epoch": 0.12623606143488322,
"grad_norm": 0.06906846165657043,
"learning_rate": 1e-05,
"loss": 2.519,
"step": 75
},
{
"epoch": 0.12791920892068168,
"grad_norm": 0.07301697880029678,
"learning_rate": 1e-05,
"loss": 2.5537,
"step": 76
},
{
"epoch": 0.12960235640648013,
"grad_norm": 0.07182423770427704,
"learning_rate": 1e-05,
"loss": 2.4807,
"step": 77
},
{
"epoch": 0.13128550389227855,
"grad_norm": 0.06827539950609207,
"learning_rate": 1e-05,
"loss": 2.5796,
"step": 78
},
{
"epoch": 0.132968651378077,
"grad_norm": 0.07280007749795914,
"learning_rate": 1e-05,
"loss": 2.499,
"step": 79
},
{
"epoch": 0.13465179886387546,
"grad_norm": 0.07410164177417755,
"learning_rate": 1e-05,
"loss": 2.3418,
"step": 80
},
{
"epoch": 0.13633494634967389,
"grad_norm": 0.07245635986328125,
"learning_rate": 1e-05,
"loss": 2.4685,
"step": 81
},
{
"epoch": 0.13801809383547234,
"grad_norm": 0.06992876529693604,
"learning_rate": 1e-05,
"loss": 2.4634,
"step": 82
},
{
"epoch": 0.13970124132127076,
"grad_norm": 0.07322832196950912,
"learning_rate": 1e-05,
"loss": 2.4949,
"step": 83
},
{
"epoch": 0.14138438880706922,
"grad_norm": 0.06528163701295853,
"learning_rate": 1e-05,
"loss": 2.3982,
"step": 84
},
{
"epoch": 0.14306753629286767,
"grad_norm": 0.06972632557153702,
"learning_rate": 1e-05,
"loss": 2.4268,
"step": 85
},
{
"epoch": 0.1447506837786661,
"grad_norm": 0.062493499368429184,
"learning_rate": 1e-05,
"loss": 2.4309,
"step": 86
},
{
"epoch": 0.14643383126446455,
"grad_norm": 0.07086165249347687,
"learning_rate": 1e-05,
"loss": 2.4373,
"step": 87
},
{
"epoch": 0.148116978750263,
"grad_norm": 0.06631726026535034,
"learning_rate": 1e-05,
"loss": 2.4141,
"step": 88
},
{
"epoch": 0.14980012623606143,
"grad_norm": 0.07114582508802414,
"learning_rate": 1e-05,
"loss": 2.3546,
"step": 89
},
{
"epoch": 0.15148327372185988,
"grad_norm": 0.06932078301906586,
"learning_rate": 1e-05,
"loss": 2.4758,
"step": 90
},
{
"epoch": 0.15316642120765833,
"grad_norm": 0.06153389438986778,
"learning_rate": 1e-05,
"loss": 2.481,
"step": 91
},
{
"epoch": 0.15484956869345676,
"grad_norm": 0.06216192990541458,
"learning_rate": 1e-05,
"loss": 2.4421,
"step": 92
},
{
"epoch": 0.1565327161792552,
"grad_norm": 0.06554314494132996,
"learning_rate": 1e-05,
"loss": 2.3008,
"step": 93
},
{
"epoch": 0.15821586366505366,
"grad_norm": 0.06210967153310776,
"learning_rate": 1e-05,
"loss": 2.2554,
"step": 94
},
{
"epoch": 0.1598990111508521,
"grad_norm": 0.06851295381784439,
"learning_rate": 1e-05,
"loss": 2.5356,
"step": 95
},
{
"epoch": 0.16158215863665054,
"grad_norm": 0.06121644005179405,
"learning_rate": 1e-05,
"loss": 2.4299,
"step": 96
},
{
"epoch": 0.16326530612244897,
"grad_norm": 0.06593657284975052,
"learning_rate": 1e-05,
"loss": 2.3811,
"step": 97
},
{
"epoch": 0.16494845360824742,
"grad_norm": 0.06456276774406433,
"learning_rate": 1e-05,
"loss": 2.3574,
"step": 98
},
{
"epoch": 0.16663160109404587,
"grad_norm": 0.061866894364356995,
"learning_rate": 1e-05,
"loss": 2.4758,
"step": 99
},
{
"epoch": 0.1683147485798443,
"grad_norm": 0.058500371873378754,
"learning_rate": 1e-05,
"loss": 2.4133,
"step": 100
},
{
"epoch": 0.16999789606564275,
"grad_norm": 0.06366603821516037,
"learning_rate": 1e-05,
"loss": 2.3328,
"step": 101
},
{
"epoch": 0.1716810435514412,
"grad_norm": 0.061924271285533905,
"learning_rate": 1e-05,
"loss": 2.4047,
"step": 102
},
{
"epoch": 0.17336419103723963,
"grad_norm": 0.057471342384815216,
"learning_rate": 1e-05,
"loss": 2.4333,
"step": 103
},
{
"epoch": 0.17504733852303808,
"grad_norm": 0.05482906475663185,
"learning_rate": 1e-05,
"loss": 2.3499,
"step": 104
},
{
"epoch": 0.17673048600883653,
"grad_norm": 0.056116051971912384,
"learning_rate": 1e-05,
"loss": 2.4653,
"step": 105
},
{
"epoch": 0.17841363349463496,
"grad_norm": 0.052277661859989166,
"learning_rate": 1e-05,
"loss": 2.4653,
"step": 106
},
{
"epoch": 0.1800967809804334,
"grad_norm": 0.06346592307090759,
"learning_rate": 1e-05,
"loss": 2.3549,
"step": 107
},
{
"epoch": 0.18177992846623187,
"grad_norm": 0.06070290133357048,
"learning_rate": 1e-05,
"loss": 2.2886,
"step": 108
},
{
"epoch": 0.1834630759520303,
"grad_norm": 0.055994004011154175,
"learning_rate": 1e-05,
"loss": 2.4692,
"step": 109
},
{
"epoch": 0.18514622343782874,
"grad_norm": 0.05782800912857056,
"learning_rate": 1e-05,
"loss": 2.3303,
"step": 110
},
{
"epoch": 0.18682937092362717,
"grad_norm": 0.05491410568356514,
"learning_rate": 1e-05,
"loss": 2.47,
"step": 111
},
{
"epoch": 0.18851251840942562,
"grad_norm": 0.060252465307712555,
"learning_rate": 1e-05,
"loss": 2.5464,
"step": 112
},
{
"epoch": 0.19019566589522408,
"grad_norm": 0.05614893510937691,
"learning_rate": 1e-05,
"loss": 2.3457,
"step": 113
},
{
"epoch": 0.1918788133810225,
"grad_norm": 0.051146939396858215,
"learning_rate": 1e-05,
"loss": 2.3918,
"step": 114
},
{
"epoch": 0.19356196086682095,
"grad_norm": 0.05474052205681801,
"learning_rate": 1e-05,
"loss": 2.3689,
"step": 115
},
{
"epoch": 0.1952451083526194,
"grad_norm": 0.052064936608076096,
"learning_rate": 1e-05,
"loss": 2.5073,
"step": 116
},
{
"epoch": 0.19692825583841783,
"grad_norm": 0.06184034049510956,
"learning_rate": 1e-05,
"loss": 2.4248,
"step": 117
},
{
"epoch": 0.19861140332421628,
"grad_norm": 0.05613533779978752,
"learning_rate": 1e-05,
"loss": 2.5742,
"step": 118
},
{
"epoch": 0.20029455081001474,
"grad_norm": 0.05547456443309784,
"learning_rate": 1e-05,
"loss": 2.3884,
"step": 119
},
{
"epoch": 0.20197769829581316,
"grad_norm": 0.05933033674955368,
"learning_rate": 1e-05,
"loss": 2.45,
"step": 120
},
{
"epoch": 0.20366084578161162,
"grad_norm": 0.058600571006536484,
"learning_rate": 1e-05,
"loss": 2.3875,
"step": 121
},
{
"epoch": 0.20534399326741007,
"grad_norm": 0.0554657019674778,
"learning_rate": 1e-05,
"loss": 2.3215,
"step": 122
},
{
"epoch": 0.2070271407532085,
"grad_norm": 0.05604475364089012,
"learning_rate": 1e-05,
"loss": 2.3329,
"step": 123
},
{
"epoch": 0.20871028823900695,
"grad_norm": 0.06094202771782875,
"learning_rate": 1e-05,
"loss": 2.4177,
"step": 124
},
{
"epoch": 0.2103934357248054,
"grad_norm": 0.05517999082803726,
"learning_rate": 1e-05,
"loss": 2.3247,
"step": 125
},
{
"epoch": 0.21207658321060383,
"grad_norm": 0.05678452178835869,
"learning_rate": 1e-05,
"loss": 2.3481,
"step": 126
},
{
"epoch": 0.21375973069640228,
"grad_norm": 0.05295870825648308,
"learning_rate": 1e-05,
"loss": 2.3694,
"step": 127
},
{
"epoch": 0.2154428781822007,
"grad_norm": 0.05118125304579735,
"learning_rate": 1e-05,
"loss": 2.4102,
"step": 128
},
{
"epoch": 0.21712602566799916,
"grad_norm": 0.05659961327910423,
"learning_rate": 1e-05,
"loss": 2.3104,
"step": 129
},
{
"epoch": 0.2188091731537976,
"grad_norm": 0.05049075558781624,
"learning_rate": 1e-05,
"loss": 2.4949,
"step": 130
},
{
"epoch": 0.22049232063959603,
"grad_norm": 0.05323097109794617,
"learning_rate": 1e-05,
"loss": 2.323,
"step": 131
},
{
"epoch": 0.2221754681253945,
"grad_norm": 0.05309610068798065,
"learning_rate": 1e-05,
"loss": 2.5203,
"step": 132
},
{
"epoch": 0.22385861561119294,
"grad_norm": 0.05474167317152023,
"learning_rate": 1e-05,
"loss": 2.408,
"step": 133
},
{
"epoch": 0.22554176309699137,
"grad_norm": 0.056433092802762985,
"learning_rate": 1e-05,
"loss": 2.3779,
"step": 134
},
{
"epoch": 0.22722491058278982,
"grad_norm": 0.047424182295799255,
"learning_rate": 1e-05,
"loss": 2.45,
"step": 135
},
{
"epoch": 0.22890805806858827,
"grad_norm": 0.05422671511769295,
"learning_rate": 1e-05,
"loss": 2.3397,
"step": 136
},
{
"epoch": 0.2305912055543867,
"grad_norm": 0.05421329662203789,
"learning_rate": 1e-05,
"loss": 2.3779,
"step": 137
},
{
"epoch": 0.23227435304018515,
"grad_norm": 0.057494040578603745,
"learning_rate": 1e-05,
"loss": 2.4509,
"step": 138
},
{
"epoch": 0.2339575005259836,
"grad_norm": 0.0516960434615612,
"learning_rate": 1e-05,
"loss": 2.3647,
"step": 139
},
{
"epoch": 0.23564064801178203,
"grad_norm": 0.049899645149707794,
"learning_rate": 1e-05,
"loss": 2.4844,
"step": 140
},
{
"epoch": 0.23732379549758048,
"grad_norm": 0.05162065476179123,
"learning_rate": 1e-05,
"loss": 2.3613,
"step": 141
},
{
"epoch": 0.2390069429833789,
"grad_norm": 0.05812832713127136,
"learning_rate": 1e-05,
"loss": 2.4548,
"step": 142
},
{
"epoch": 0.24069009046917736,
"grad_norm": 0.04910556599497795,
"learning_rate": 1e-05,
"loss": 2.3274,
"step": 143
},
{
"epoch": 0.2423732379549758,
"grad_norm": 0.05346587672829628,
"learning_rate": 1e-05,
"loss": 2.325,
"step": 144
},
{
"epoch": 0.24405638544077424,
"grad_norm": 0.0495002381503582,
"learning_rate": 1e-05,
"loss": 2.4131,
"step": 145
},
{
"epoch": 0.2457395329265727,
"grad_norm": 0.05076875165104866,
"learning_rate": 1e-05,
"loss": 2.3887,
"step": 146
},
{
"epoch": 0.24742268041237114,
"grad_norm": 0.050955574959516525,
"learning_rate": 1e-05,
"loss": 2.4517,
"step": 147
},
{
"epoch": 0.24910582789816957,
"grad_norm": 0.05082906410098076,
"learning_rate": 1e-05,
"loss": 2.3401,
"step": 148
},
{
"epoch": 0.250788975383968,
"grad_norm": 0.052096717059612274,
"learning_rate": 1e-05,
"loss": 2.3218,
"step": 149
},
{
"epoch": 0.25247212286976645,
"grad_norm": 0.052378151565790176,
"learning_rate": 1e-05,
"loss": 2.4246,
"step": 150
},
{
"epoch": 0.2541552703555649,
"grad_norm": 0.04881056025624275,
"learning_rate": 1e-05,
"loss": 2.3435,
"step": 151
},
{
"epoch": 0.25583841784136335,
"grad_norm": 0.05233067274093628,
"learning_rate": 1e-05,
"loss": 2.4761,
"step": 152
},
{
"epoch": 0.2575215653271618,
"grad_norm": 0.05231297388672829,
"learning_rate": 1e-05,
"loss": 2.4065,
"step": 153
},
{
"epoch": 0.25920471281296026,
"grad_norm": 0.04649129509925842,
"learning_rate": 1e-05,
"loss": 2.4175,
"step": 154
},
{
"epoch": 0.26088786029875866,
"grad_norm": 0.05354660376906395,
"learning_rate": 1e-05,
"loss": 2.4731,
"step": 155
},
{
"epoch": 0.2625710077845571,
"grad_norm": 0.05071151629090309,
"learning_rate": 1e-05,
"loss": 2.4421,
"step": 156
},
{
"epoch": 0.26425415527035556,
"grad_norm": 0.04953297600150108,
"learning_rate": 1e-05,
"loss": 2.3134,
"step": 157
},
{
"epoch": 0.265937302756154,
"grad_norm": 0.051142722368240356,
"learning_rate": 1e-05,
"loss": 2.3335,
"step": 158
},
{
"epoch": 0.26762045024195247,
"grad_norm": 0.05187085270881653,
"learning_rate": 1e-05,
"loss": 2.4387,
"step": 159
},
{
"epoch": 0.2693035977277509,
"grad_norm": 0.04968629032373428,
"learning_rate": 1e-05,
"loss": 2.4905,
"step": 160
},
{
"epoch": 0.2709867452135493,
"grad_norm": 0.053009629249572754,
"learning_rate": 1e-05,
"loss": 2.4441,
"step": 161
},
{
"epoch": 0.27266989269934777,
"grad_norm": 0.04917874187231064,
"learning_rate": 1e-05,
"loss": 2.4763,
"step": 162
},
{
"epoch": 0.2743530401851462,
"grad_norm": 0.048884451389312744,
"learning_rate": 1e-05,
"loss": 2.4248,
"step": 163
},
{
"epoch": 0.2760361876709447,
"grad_norm": 0.049946676939725876,
"learning_rate": 1e-05,
"loss": 2.5173,
"step": 164
},
{
"epoch": 0.27771933515674313,
"grad_norm": 0.052534863352775574,
"learning_rate": 1e-05,
"loss": 2.4558,
"step": 165
},
{
"epoch": 0.2794024826425415,
"grad_norm": 0.05162844434380531,
"learning_rate": 1e-05,
"loss": 2.405,
"step": 166
},
{
"epoch": 0.28108563012834,
"grad_norm": 0.049985259771347046,
"learning_rate": 1e-05,
"loss": 2.3542,
"step": 167
},
{
"epoch": 0.28276877761413843,
"grad_norm": 0.05239354074001312,
"learning_rate": 1e-05,
"loss": 2.3721,
"step": 168
},
{
"epoch": 0.2844519250999369,
"grad_norm": 0.05592744052410126,
"learning_rate": 1e-05,
"loss": 2.2701,
"step": 169
},
{
"epoch": 0.28613507258573534,
"grad_norm": 0.052739113569259644,
"learning_rate": 1e-05,
"loss": 2.4216,
"step": 170
},
{
"epoch": 0.2878182200715338,
"grad_norm": 0.04806948080658913,
"learning_rate": 1e-05,
"loss": 2.3884,
"step": 171
},
{
"epoch": 0.2895013675573322,
"grad_norm": 0.04990949481725693,
"learning_rate": 1e-05,
"loss": 2.4419,
"step": 172
},
{
"epoch": 0.29118451504313064,
"grad_norm": 0.050067439675331116,
"learning_rate": 1e-05,
"loss": 2.4331,
"step": 173
},
{
"epoch": 0.2928676625289291,
"grad_norm": 0.0507354810833931,
"learning_rate": 1e-05,
"loss": 2.406,
"step": 174
},
{
"epoch": 0.29455081001472755,
"grad_norm": 0.0538686104118824,
"learning_rate": 1e-05,
"loss": 2.4182,
"step": 175
},
{
"epoch": 0.296233957500526,
"grad_norm": 0.05205219238996506,
"learning_rate": 1e-05,
"loss": 2.3401,
"step": 176
},
{
"epoch": 0.2979171049863244,
"grad_norm": 0.04672086611390114,
"learning_rate": 1e-05,
"loss": 2.3149,
"step": 177
},
{
"epoch": 0.29960025247212285,
"grad_norm": 0.051963068544864655,
"learning_rate": 1e-05,
"loss": 2.2537,
"step": 178
},
{
"epoch": 0.3012833999579213,
"grad_norm": 0.053639005869627,
"learning_rate": 1e-05,
"loss": 2.4353,
"step": 179
},
{
"epoch": 0.30296654744371976,
"grad_norm": 0.05326982960104942,
"learning_rate": 1e-05,
"loss": 2.334,
"step": 180
},
{
"epoch": 0.3046496949295182,
"grad_norm": 0.05361334979534149,
"learning_rate": 1e-05,
"loss": 2.4224,
"step": 181
},
{
"epoch": 0.30633284241531666,
"grad_norm": 0.05790587514638901,
"learning_rate": 1e-05,
"loss": 2.334,
"step": 182
},
{
"epoch": 0.30801598990111506,
"grad_norm": 0.04790763929486275,
"learning_rate": 1e-05,
"loss": 2.5073,
"step": 183
},
{
"epoch": 0.3096991373869135,
"grad_norm": 0.054103124886751175,
"learning_rate": 1e-05,
"loss": 2.3483,
"step": 184
},
{
"epoch": 0.31138228487271197,
"grad_norm": 0.05902162939310074,
"learning_rate": 1e-05,
"loss": 2.3301,
"step": 185
},
{
"epoch": 0.3130654323585104,
"grad_norm": 0.04853544384241104,
"learning_rate": 1e-05,
"loss": 2.5566,
"step": 186
},
{
"epoch": 0.3147485798443089,
"grad_norm": 0.055288165807724,
"learning_rate": 1e-05,
"loss": 2.2903,
"step": 187
},
{
"epoch": 0.3164317273301073,
"grad_norm": 0.05180734023451805,
"learning_rate": 1e-05,
"loss": 2.4285,
"step": 188
},
{
"epoch": 0.3181148748159057,
"grad_norm": 0.04889997839927673,
"learning_rate": 1e-05,
"loss": 2.2542,
"step": 189
},
{
"epoch": 0.3197980223017042,
"grad_norm": 0.051011502742767334,
"learning_rate": 1e-05,
"loss": 2.2893,
"step": 190
},
{
"epoch": 0.32148116978750263,
"grad_norm": 0.04864371567964554,
"learning_rate": 1e-05,
"loss": 2.5225,
"step": 191
},
{
"epoch": 0.3231643172733011,
"grad_norm": 0.05374041944742203,
"learning_rate": 1e-05,
"loss": 2.4504,
"step": 192
},
{
"epoch": 0.32484746475909954,
"grad_norm": 0.05158041790127754,
"learning_rate": 1e-05,
"loss": 2.4683,
"step": 193
},
{
"epoch": 0.32653061224489793,
"grad_norm": 0.05630083382129669,
"learning_rate": 1e-05,
"loss": 2.2415,
"step": 194
},
{
"epoch": 0.3282137597306964,
"grad_norm": 0.05439196154475212,
"learning_rate": 1e-05,
"loss": 2.3684,
"step": 195
},
{
"epoch": 0.32989690721649484,
"grad_norm": 0.05023415759205818,
"learning_rate": 1e-05,
"loss": 2.415,
"step": 196
},
{
"epoch": 0.3315800547022933,
"grad_norm": 0.05531445890665054,
"learning_rate": 1e-05,
"loss": 2.4626,
"step": 197
},
{
"epoch": 0.33326320218809175,
"grad_norm": 0.05087656155228615,
"learning_rate": 1e-05,
"loss": 2.3936,
"step": 198
},
{
"epoch": 0.3349463496738902,
"grad_norm": 0.05231088399887085,
"learning_rate": 1e-05,
"loss": 2.3779,
"step": 199
},
{
"epoch": 0.3366294971596886,
"grad_norm": 0.0514984093606472,
"learning_rate": 1e-05,
"loss": 2.3967,
"step": 200
},
{
"epoch": 0.33831264464548705,
"grad_norm": 0.05334719642996788,
"learning_rate": 1e-05,
"loss": 2.4604,
"step": 201
},
{
"epoch": 0.3399957921312855,
"grad_norm": 0.054843124002218246,
"learning_rate": 1e-05,
"loss": 2.3538,
"step": 202
},
{
"epoch": 0.34167893961708395,
"grad_norm": 0.04888272285461426,
"learning_rate": 1e-05,
"loss": 2.4844,
"step": 203
},
{
"epoch": 0.3433620871028824,
"grad_norm": 0.054122187197208405,
"learning_rate": 1e-05,
"loss": 2.3291,
"step": 204
},
{
"epoch": 0.34504523458868086,
"grad_norm": 0.054561201483011246,
"learning_rate": 1e-05,
"loss": 2.3218,
"step": 205
},
{
"epoch": 0.34672838207447926,
"grad_norm": 0.04919834062457085,
"learning_rate": 1e-05,
"loss": 2.4478,
"step": 206
},
{
"epoch": 0.3484115295602777,
"grad_norm": 0.050551943480968475,
"learning_rate": 1e-05,
"loss": 2.3755,
"step": 207
},
{
"epoch": 0.35009467704607616,
"grad_norm": 0.05242514982819557,
"learning_rate": 1e-05,
"loss": 2.3922,
"step": 208
},
{
"epoch": 0.3517778245318746,
"grad_norm": 0.06077054515480995,
"learning_rate": 1e-05,
"loss": 2.3218,
"step": 209
},
{
"epoch": 0.35346097201767307,
"grad_norm": 0.061367545276880264,
"learning_rate": 1e-05,
"loss": 2.2957,
"step": 210
},
{
"epoch": 0.35514411950347147,
"grad_norm": 0.0511772483587265,
"learning_rate": 1e-05,
"loss": 2.374,
"step": 211
},
{
"epoch": 0.3568272669892699,
"grad_norm": 0.0496203638613224,
"learning_rate": 1e-05,
"loss": 2.4182,
"step": 212
},
{
"epoch": 0.3585104144750684,
"grad_norm": 0.061339233070611954,
"learning_rate": 1e-05,
"loss": 2.406,
"step": 213
},
{
"epoch": 0.3601935619608668,
"grad_norm": 0.052460432052612305,
"learning_rate": 1e-05,
"loss": 2.4309,
"step": 214
},
{
"epoch": 0.3618767094466653,
"grad_norm": 0.055436089634895325,
"learning_rate": 1e-05,
"loss": 2.4141,
"step": 215
},
{
"epoch": 0.36355985693246373,
"grad_norm": 0.05396036058664322,
"learning_rate": 1e-05,
"loss": 2.2705,
"step": 216
},
{
"epoch": 0.36524300441826213,
"grad_norm": 0.04853086173534393,
"learning_rate": 1e-05,
"loss": 2.4473,
"step": 217
},
{
"epoch": 0.3669261519040606,
"grad_norm": 0.051015399396419525,
"learning_rate": 1e-05,
"loss": 2.5115,
"step": 218
},
{
"epoch": 0.36860929938985904,
"grad_norm": 0.05526035279035568,
"learning_rate": 1e-05,
"loss": 2.3123,
"step": 219
},
{
"epoch": 0.3702924468756575,
"grad_norm": 0.056169234216213226,
"learning_rate": 1e-05,
"loss": 2.3447,
"step": 220
},
{
"epoch": 0.37197559436145594,
"grad_norm": 0.05238133668899536,
"learning_rate": 1e-05,
"loss": 2.26,
"step": 221
},
{
"epoch": 0.37365874184725434,
"grad_norm": 0.05587685480713844,
"learning_rate": 1e-05,
"loss": 2.3083,
"step": 222
},
{
"epoch": 0.3753418893330528,
"grad_norm": 0.050364553928375244,
"learning_rate": 1e-05,
"loss": 2.3459,
"step": 223
},
{
"epoch": 0.37702503681885124,
"grad_norm": 0.0506574809551239,
"learning_rate": 1e-05,
"loss": 2.4246,
"step": 224
},
{
"epoch": 0.3787081843046497,
"grad_norm": 0.05842865630984306,
"learning_rate": 1e-05,
"loss": 2.2617,
"step": 225
},
{
"epoch": 0.38039133179044815,
"grad_norm": 0.05097496882081032,
"learning_rate": 1e-05,
"loss": 2.52,
"step": 226
},
{
"epoch": 0.3820744792762466,
"grad_norm": 0.05665278434753418,
"learning_rate": 1e-05,
"loss": 2.2715,
"step": 227
},
{
"epoch": 0.383757626762045,
"grad_norm": 0.053350359201431274,
"learning_rate": 1e-05,
"loss": 2.3101,
"step": 228
},
{
"epoch": 0.38544077424784345,
"grad_norm": 0.05481604114174843,
"learning_rate": 1e-05,
"loss": 2.3347,
"step": 229
},
{
"epoch": 0.3871239217336419,
"grad_norm": 0.06036606431007385,
"learning_rate": 1e-05,
"loss": 2.2991,
"step": 230
},
{
"epoch": 0.38880706921944036,
"grad_norm": 0.0606355145573616,
"learning_rate": 1e-05,
"loss": 2.4226,
"step": 231
},
{
"epoch": 0.3904902167052388,
"grad_norm": 0.052770137786865234,
"learning_rate": 1e-05,
"loss": 2.4539,
"step": 232
},
{
"epoch": 0.39217336419103727,
"grad_norm": 0.050006203353405,
"learning_rate": 1e-05,
"loss": 2.3477,
"step": 233
},
{
"epoch": 0.39385651167683566,
"grad_norm": 0.05640649050474167,
"learning_rate": 1e-05,
"loss": 2.3123,
"step": 234
},
{
"epoch": 0.3955396591626341,
"grad_norm": 0.050969429314136505,
"learning_rate": 1e-05,
"loss": 2.4534,
"step": 235
},
{
"epoch": 0.39722280664843257,
"grad_norm": 0.05676101893186569,
"learning_rate": 1e-05,
"loss": 2.3481,
"step": 236
},
{
"epoch": 0.398905954134231,
"grad_norm": 0.05844707787036896,
"learning_rate": 1e-05,
"loss": 2.3638,
"step": 237
},
{
"epoch": 0.4005891016200295,
"grad_norm": 0.053074926137924194,
"learning_rate": 1e-05,
"loss": 2.3904,
"step": 238
},
{
"epoch": 0.4022722491058279,
"grad_norm": 0.04979414492845535,
"learning_rate": 1e-05,
"loss": 2.3855,
"step": 239
},
{
"epoch": 0.4039553965916263,
"grad_norm": 0.05607665330171585,
"learning_rate": 1e-05,
"loss": 2.3569,
"step": 240
},
{
"epoch": 0.4056385440774248,
"grad_norm": 0.05964501202106476,
"learning_rate": 1e-05,
"loss": 2.3459,
"step": 241
},
{
"epoch": 0.40732169156322323,
"grad_norm": 0.05849093198776245,
"learning_rate": 1e-05,
"loss": 2.3213,
"step": 242
},
{
"epoch": 0.4090048390490217,
"grad_norm": 0.053846072405576706,
"learning_rate": 1e-05,
"loss": 2.4436,
"step": 243
},
{
"epoch": 0.41068798653482014,
"grad_norm": 0.054448988288640976,
"learning_rate": 1e-05,
"loss": 2.3716,
"step": 244
},
{
"epoch": 0.41237113402061853,
"grad_norm": 0.05229583755135536,
"learning_rate": 1e-05,
"loss": 2.4099,
"step": 245
},
{
"epoch": 0.414054281506417,
"grad_norm": 0.05479966476559639,
"learning_rate": 1e-05,
"loss": 2.4026,
"step": 246
},
{
"epoch": 0.41573742899221544,
"grad_norm": 0.061799049377441406,
"learning_rate": 1e-05,
"loss": 2.4072,
"step": 247
},
{
"epoch": 0.4174205764780139,
"grad_norm": 0.061452727764844894,
"learning_rate": 1e-05,
"loss": 2.2833,
"step": 248
},
{
"epoch": 0.41910372396381235,
"grad_norm": 0.05868072435259819,
"learning_rate": 1e-05,
"loss": 2.3833,
"step": 249
},
{
"epoch": 0.4207868714496108,
"grad_norm": 0.05926290899515152,
"learning_rate": 1e-05,
"loss": 2.3645,
"step": 250
},
{
"epoch": 0.4224700189354092,
"grad_norm": 0.058858342468738556,
"learning_rate": 1e-05,
"loss": 2.3152,
"step": 251
},
{
"epoch": 0.42415316642120765,
"grad_norm": 0.058599065989255905,
"learning_rate": 1e-05,
"loss": 2.2827,
"step": 252
},
{
"epoch": 0.4258363139070061,
"grad_norm": 0.060381706804037094,
"learning_rate": 1e-05,
"loss": 2.3024,
"step": 253
},
{
"epoch": 0.42751946139280456,
"grad_norm": 0.05441940575838089,
"learning_rate": 1e-05,
"loss": 2.446,
"step": 254
},
{
"epoch": 0.429202608878603,
"grad_norm": 0.05750846117734909,
"learning_rate": 1e-05,
"loss": 2.3958,
"step": 255
},
{
"epoch": 0.4308857563644014,
"grad_norm": 0.060346368700265884,
"learning_rate": 1e-05,
"loss": 2.2395,
"step": 256
},
{
"epoch": 0.43256890385019986,
"grad_norm": 0.056383710354566574,
"learning_rate": 1e-05,
"loss": 2.3518,
"step": 257
},
{
"epoch": 0.4342520513359983,
"grad_norm": 0.057746805250644684,
"learning_rate": 1e-05,
"loss": 2.2834,
"step": 258
},
{
"epoch": 0.43593519882179677,
"grad_norm": 0.051562029868364334,
"learning_rate": 1e-05,
"loss": 2.3677,
"step": 259
},
{
"epoch": 0.4376183463075952,
"grad_norm": 0.059988316148519516,
"learning_rate": 1e-05,
"loss": 2.3372,
"step": 260
},
{
"epoch": 0.43930149379339367,
"grad_norm": 0.05852155759930611,
"learning_rate": 1e-05,
"loss": 2.3875,
"step": 261
},
{
"epoch": 0.44098464127919207,
"grad_norm": 0.06629418581724167,
"learning_rate": 1e-05,
"loss": 2.4194,
"step": 262
},
{
"epoch": 0.4426677887649905,
"grad_norm": 0.061044465750455856,
"learning_rate": 1e-05,
"loss": 2.2466,
"step": 263
},
{
"epoch": 0.444350936250789,
"grad_norm": 0.056285977363586426,
"learning_rate": 1e-05,
"loss": 2.3105,
"step": 264
},
{
"epoch": 0.44603408373658743,
"grad_norm": 0.06135227158665657,
"learning_rate": 1e-05,
"loss": 2.3853,
"step": 265
},
{
"epoch": 0.4477172312223859,
"grad_norm": 0.05644640699028969,
"learning_rate": 1e-05,
"loss": 2.3888,
"step": 266
},
{
"epoch": 0.4494003787081843,
"grad_norm": 0.06326981633901596,
"learning_rate": 1e-05,
"loss": 2.3132,
"step": 267
},
{
"epoch": 0.45108352619398273,
"grad_norm": 0.05710430070757866,
"learning_rate": 1e-05,
"loss": 2.365,
"step": 268
},
{
"epoch": 0.4527666736797812,
"grad_norm": 0.05607946217060089,
"learning_rate": 1e-05,
"loss": 2.4648,
"step": 269
},
{
"epoch": 0.45444982116557964,
"grad_norm": 0.057825781404972076,
"learning_rate": 1e-05,
"loss": 2.4189,
"step": 270
},
{
"epoch": 0.4561329686513781,
"grad_norm": 0.06380680948495865,
"learning_rate": 1e-05,
"loss": 2.3188,
"step": 271
},
{
"epoch": 0.45781611613717654,
"grad_norm": 0.06377760320901871,
"learning_rate": 1e-05,
"loss": 2.2896,
"step": 272
},
{
"epoch": 0.45949926362297494,
"grad_norm": 0.06210333853960037,
"learning_rate": 1e-05,
"loss": 2.3663,
"step": 273
},
{
"epoch": 0.4611824111087734,
"grad_norm": 0.06039275974035263,
"learning_rate": 1e-05,
"loss": 2.408,
"step": 274
},
{
"epoch": 0.46286555859457185,
"grad_norm": 0.05442138388752937,
"learning_rate": 1e-05,
"loss": 2.3843,
"step": 275
},
{
"epoch": 0.4645487060803703,
"grad_norm": 0.06208937615156174,
"learning_rate": 1e-05,
"loss": 2.4355,
"step": 276
},
{
"epoch": 0.46623185356616875,
"grad_norm": 0.0619891993701458,
"learning_rate": 1e-05,
"loss": 2.3196,
"step": 277
},
{
"epoch": 0.4679150010519672,
"grad_norm": 0.059192296117544174,
"learning_rate": 1e-05,
"loss": 2.3237,
"step": 278
},
{
"epoch": 0.4695981485377656,
"grad_norm": 0.06284468621015549,
"learning_rate": 1e-05,
"loss": 2.3694,
"step": 279
},
{
"epoch": 0.47128129602356406,
"grad_norm": 0.06121189519762993,
"learning_rate": 1e-05,
"loss": 2.3606,
"step": 280
},
{
"epoch": 0.4729644435093625,
"grad_norm": 0.061919402331113815,
"learning_rate": 1e-05,
"loss": 2.3381,
"step": 281
},
{
"epoch": 0.47464759099516096,
"grad_norm": 0.0676443800330162,
"learning_rate": 1e-05,
"loss": 2.3624,
"step": 282
},
{
"epoch": 0.4763307384809594,
"grad_norm": 0.060140665620565414,
"learning_rate": 1e-05,
"loss": 2.4541,
"step": 283
},
{
"epoch": 0.4780138859667578,
"grad_norm": 0.062285441905260086,
"learning_rate": 1e-05,
"loss": 2.323,
"step": 284
},
{
"epoch": 0.47969703345255627,
"grad_norm": 0.06063227355480194,
"learning_rate": 1e-05,
"loss": 2.3596,
"step": 285
},
{
"epoch": 0.4813801809383547,
"grad_norm": 0.05906851589679718,
"learning_rate": 1e-05,
"loss": 2.458,
"step": 286
},
{
"epoch": 0.48306332842415317,
"grad_norm": 0.05862203240394592,
"learning_rate": 1e-05,
"loss": 2.291,
"step": 287
},
{
"epoch": 0.4847464759099516,
"grad_norm": 0.0629325732588768,
"learning_rate": 1e-05,
"loss": 2.2634,
"step": 288
},
{
"epoch": 0.4864296233957501,
"grad_norm": 0.06464157998561859,
"learning_rate": 1e-05,
"loss": 2.2531,
"step": 289
},
{
"epoch": 0.4881127708815485,
"grad_norm": 0.0547555610537529,
"learning_rate": 1e-05,
"loss": 2.5339,
"step": 290
},
{
"epoch": 0.4897959183673469,
"grad_norm": 0.0606168657541275,
"learning_rate": 1e-05,
"loss": 2.2886,
"step": 291
},
{
"epoch": 0.4914790658531454,
"grad_norm": 0.058814577758312225,
"learning_rate": 1e-05,
"loss": 2.3337,
"step": 292
},
{
"epoch": 0.49316221333894383,
"grad_norm": 0.0691385492682457,
"learning_rate": 1e-05,
"loss": 2.2904,
"step": 293
},
{
"epoch": 0.4948453608247423,
"grad_norm": 0.06522157788276672,
"learning_rate": 1e-05,
"loss": 2.469,
"step": 294
},
{
"epoch": 0.4965285083105407,
"grad_norm": 0.05957287177443504,
"learning_rate": 1e-05,
"loss": 2.4095,
"step": 295
},
{
"epoch": 0.49821165579633914,
"grad_norm": 0.06277060508728027,
"learning_rate": 1e-05,
"loss": 2.4697,
"step": 296
},
{
"epoch": 0.4998948032821376,
"grad_norm": 0.06802426278591156,
"learning_rate": 1e-05,
"loss": 2.2517,
"step": 297
},
{
"epoch": 0.501577950767936,
"grad_norm": 0.06365792453289032,
"learning_rate": 1e-05,
"loss": 2.2942,
"step": 298
},
{
"epoch": 0.5032610982537344,
"grad_norm": 0.06624794751405716,
"learning_rate": 1e-05,
"loss": 2.283,
"step": 299
},
{
"epoch": 0.5049442457395329,
"grad_norm": 0.05979595705866814,
"learning_rate": 1e-05,
"loss": 2.4387,
"step": 300
},
{
"epoch": 0.5066273932253313,
"grad_norm": 0.06187634915113449,
"learning_rate": 1e-05,
"loss": 2.4205,
"step": 301
},
{
"epoch": 0.5083105407111298,
"grad_norm": 0.06389462947845459,
"learning_rate": 1e-05,
"loss": 2.2775,
"step": 302
},
{
"epoch": 0.5099936881969283,
"grad_norm": 0.05831071361899376,
"learning_rate": 1e-05,
"loss": 2.3892,
"step": 303
},
{
"epoch": 0.5116768356827267,
"grad_norm": 0.06568494439125061,
"learning_rate": 1e-05,
"loss": 2.3087,
"step": 304
},
{
"epoch": 0.5133599831685252,
"grad_norm": 0.062109317630529404,
"learning_rate": 1e-05,
"loss": 2.3268,
"step": 305
},
{
"epoch": 0.5150431306543236,
"grad_norm": 0.061168327927589417,
"learning_rate": 1e-05,
"loss": 2.3093,
"step": 306
},
{
"epoch": 0.5167262781401221,
"grad_norm": 0.061159648001194,
"learning_rate": 1e-05,
"loss": 2.3315,
"step": 307
},
{
"epoch": 0.5184094256259205,
"grad_norm": 0.06269169598817825,
"learning_rate": 1e-05,
"loss": 2.3442,
"step": 308
},
{
"epoch": 0.520092573111719,
"grad_norm": 0.06711502373218536,
"learning_rate": 1e-05,
"loss": 2.2008,
"step": 309
},
{
"epoch": 0.5217757205975173,
"grad_norm": 0.0663105845451355,
"learning_rate": 1e-05,
"loss": 2.3502,
"step": 310
},
{
"epoch": 0.5234588680833158,
"grad_norm": 0.06040646880865097,
"learning_rate": 1e-05,
"loss": 2.3414,
"step": 311
},
{
"epoch": 0.5251420155691142,
"grad_norm": 0.06823603063821793,
"learning_rate": 1e-05,
"loss": 2.3392,
"step": 312
},
{
"epoch": 0.5268251630549127,
"grad_norm": 0.05944176763296127,
"learning_rate": 1e-05,
"loss": 2.3193,
"step": 313
},
{
"epoch": 0.5285083105407111,
"grad_norm": 0.06610157340765,
"learning_rate": 1e-05,
"loss": 2.2288,
"step": 314
},
{
"epoch": 0.5301914580265096,
"grad_norm": 0.06880299746990204,
"learning_rate": 1e-05,
"loss": 2.3529,
"step": 315
},
{
"epoch": 0.531874605512308,
"grad_norm": 0.06061836704611778,
"learning_rate": 1e-05,
"loss": 2.3533,
"step": 316
},
{
"epoch": 0.5335577529981065,
"grad_norm": 0.06552371382713318,
"learning_rate": 1e-05,
"loss": 2.3579,
"step": 317
},
{
"epoch": 0.5352409004839049,
"grad_norm": 0.06967922300100327,
"learning_rate": 1e-05,
"loss": 2.2983,
"step": 318
},
{
"epoch": 0.5369240479697034,
"grad_norm": 0.06997574120759964,
"learning_rate": 1e-05,
"loss": 2.355,
"step": 319
},
{
"epoch": 0.5386071954555018,
"grad_norm": 0.0654403418302536,
"learning_rate": 1e-05,
"loss": 2.4258,
"step": 320
},
{
"epoch": 0.5402903429413002,
"grad_norm": 0.06031208485364914,
"learning_rate": 1e-05,
"loss": 2.4011,
"step": 321
},
{
"epoch": 0.5419734904270986,
"grad_norm": 0.06496379524469376,
"learning_rate": 1e-05,
"loss": 2.2429,
"step": 322
},
{
"epoch": 0.5436566379128971,
"grad_norm": 0.06525281816720963,
"learning_rate": 1e-05,
"loss": 2.3254,
"step": 323
},
{
"epoch": 0.5453397853986955,
"grad_norm": 0.07553514093160629,
"learning_rate": 1e-05,
"loss": 2.2953,
"step": 324
},
{
"epoch": 0.547022932884494,
"grad_norm": 0.06429509073495865,
"learning_rate": 1e-05,
"loss": 2.3319,
"step": 325
},
{
"epoch": 0.5487060803702924,
"grad_norm": 0.0657946914434433,
"learning_rate": 1e-05,
"loss": 2.3501,
"step": 326
},
{
"epoch": 0.5503892278560909,
"grad_norm": 0.06548567861318588,
"learning_rate": 1e-05,
"loss": 2.2781,
"step": 327
},
{
"epoch": 0.5520723753418894,
"grad_norm": 0.06299672275781631,
"learning_rate": 1e-05,
"loss": 2.377,
"step": 328
},
{
"epoch": 0.5537555228276878,
"grad_norm": 0.06381850689649582,
"learning_rate": 1e-05,
"loss": 2.3945,
"step": 329
},
{
"epoch": 0.5554386703134863,
"grad_norm": 0.06497140228748322,
"learning_rate": 1e-05,
"loss": 2.3496,
"step": 330
},
{
"epoch": 0.5571218177992847,
"grad_norm": 0.06588133424520493,
"learning_rate": 1e-05,
"loss": 2.3955,
"step": 331
},
{
"epoch": 0.558804965285083,
"grad_norm": 0.06468643248081207,
"learning_rate": 1e-05,
"loss": 2.2893,
"step": 332
},
{
"epoch": 0.5604881127708815,
"grad_norm": 0.07278285920619965,
"learning_rate": 1e-05,
"loss": 2.3179,
"step": 333
},
{
"epoch": 0.56217126025668,
"grad_norm": 0.06992325931787491,
"learning_rate": 1e-05,
"loss": 2.3588,
"step": 334
},
{
"epoch": 0.5638544077424784,
"grad_norm": 0.06566626578569412,
"learning_rate": 1e-05,
"loss": 2.4763,
"step": 335
},
{
"epoch": 0.5655375552282769,
"grad_norm": 0.0633927658200264,
"learning_rate": 1e-05,
"loss": 2.4685,
"step": 336
},
{
"epoch": 0.5672207027140753,
"grad_norm": 0.06903122365474701,
"learning_rate": 1e-05,
"loss": 2.311,
"step": 337
},
{
"epoch": 0.5689038501998738,
"grad_norm": 0.06421441584825516,
"learning_rate": 1e-05,
"loss": 2.3589,
"step": 338
},
{
"epoch": 0.5705869976856722,
"grad_norm": 0.07122648507356644,
"learning_rate": 1e-05,
"loss": 2.3798,
"step": 339
},
{
"epoch": 0.5722701451714707,
"grad_norm": 0.06518077105283737,
"learning_rate": 1e-05,
"loss": 2.4546,
"step": 340
},
{
"epoch": 0.5739532926572691,
"grad_norm": 0.07509720325469971,
"learning_rate": 1e-05,
"loss": 2.3341,
"step": 341
},
{
"epoch": 0.5756364401430676,
"grad_norm": 0.06559302657842636,
"learning_rate": 1e-05,
"loss": 2.3127,
"step": 342
},
{
"epoch": 0.5773195876288659,
"grad_norm": 0.06652245670557022,
"learning_rate": 1e-05,
"loss": 2.3997,
"step": 343
},
{
"epoch": 0.5790027351146644,
"grad_norm": 0.07472145557403564,
"learning_rate": 1e-05,
"loss": 2.3237,
"step": 344
},
{
"epoch": 0.5806858826004628,
"grad_norm": 0.07624109089374542,
"learning_rate": 1e-05,
"loss": 2.186,
"step": 345
},
{
"epoch": 0.5823690300862613,
"grad_norm": 0.06387084722518921,
"learning_rate": 1e-05,
"loss": 2.2717,
"step": 346
},
{
"epoch": 0.5840521775720597,
"grad_norm": 0.06857839971780777,
"learning_rate": 1e-05,
"loss": 2.3726,
"step": 347
},
{
"epoch": 0.5857353250578582,
"grad_norm": 0.06429892778396606,
"learning_rate": 1e-05,
"loss": 2.4109,
"step": 348
},
{
"epoch": 0.5874184725436566,
"grad_norm": 0.0720372200012207,
"learning_rate": 1e-05,
"loss": 2.3291,
"step": 349
},
{
"epoch": 0.5891016200294551,
"grad_norm": 0.0749678909778595,
"learning_rate": 1e-05,
"loss": 2.3369,
"step": 350
},
{
"epoch": 0.5907847675152536,
"grad_norm": 0.0645705908536911,
"learning_rate": 1e-05,
"loss": 2.3894,
"step": 351
},
{
"epoch": 0.592467915001052,
"grad_norm": 0.06680341064929962,
"learning_rate": 1e-05,
"loss": 2.3335,
"step": 352
},
{
"epoch": 0.5941510624868505,
"grad_norm": 0.07383781671524048,
"learning_rate": 1e-05,
"loss": 2.2733,
"step": 353
},
{
"epoch": 0.5958342099726488,
"grad_norm": 0.07338624447584152,
"learning_rate": 1e-05,
"loss": 2.2236,
"step": 354
},
{
"epoch": 0.5975173574584473,
"grad_norm": 0.06998410820960999,
"learning_rate": 1e-05,
"loss": 2.2552,
"step": 355
},
{
"epoch": 0.5992005049442457,
"grad_norm": 0.06697436422109604,
"learning_rate": 1e-05,
"loss": 2.4231,
"step": 356
},
{
"epoch": 0.6008836524300442,
"grad_norm": 0.06693920493125916,
"learning_rate": 1e-05,
"loss": 2.3296,
"step": 357
},
{
"epoch": 0.6025667999158426,
"grad_norm": 0.06306028366088867,
"learning_rate": 1e-05,
"loss": 2.4009,
"step": 358
},
{
"epoch": 0.6042499474016411,
"grad_norm": 0.0724472776055336,
"learning_rate": 1e-05,
"loss": 2.2986,
"step": 359
},
{
"epoch": 0.6059330948874395,
"grad_norm": 0.06711563467979431,
"learning_rate": 1e-05,
"loss": 2.3755,
"step": 360
},
{
"epoch": 0.607616242373238,
"grad_norm": 0.07287666201591492,
"learning_rate": 1e-05,
"loss": 2.325,
"step": 361
},
{
"epoch": 0.6092993898590364,
"grad_norm": 0.07494334876537323,
"learning_rate": 1e-05,
"loss": 2.2673,
"step": 362
},
{
"epoch": 0.6109825373448349,
"grad_norm": 0.07399529218673706,
"learning_rate": 1e-05,
"loss": 2.3134,
"step": 363
},
{
"epoch": 0.6126656848306333,
"grad_norm": 0.06705833226442337,
"learning_rate": 1e-05,
"loss": 2.3772,
"step": 364
},
{
"epoch": 0.6143488323164318,
"grad_norm": 0.07528689503669739,
"learning_rate": 1e-05,
"loss": 2.3872,
"step": 365
},
{
"epoch": 0.6160319798022301,
"grad_norm": 0.06814612448215485,
"learning_rate": 1e-05,
"loss": 2.2527,
"step": 366
},
{
"epoch": 0.6177151272880286,
"grad_norm": 0.06929857283830643,
"learning_rate": 1e-05,
"loss": 2.4138,
"step": 367
},
{
"epoch": 0.619398274773827,
"grad_norm": 0.07336314767599106,
"learning_rate": 1e-05,
"loss": 2.4197,
"step": 368
},
{
"epoch": 0.6210814222596255,
"grad_norm": 0.07009201496839523,
"learning_rate": 1e-05,
"loss": 2.3943,
"step": 369
},
{
"epoch": 0.6227645697454239,
"grad_norm": 0.07367721945047379,
"learning_rate": 1e-05,
"loss": 2.3044,
"step": 370
},
{
"epoch": 0.6244477172312224,
"grad_norm": 0.07029354572296143,
"learning_rate": 1e-05,
"loss": 2.3018,
"step": 371
},
{
"epoch": 0.6261308647170208,
"grad_norm": 0.07852700352668762,
"learning_rate": 1e-05,
"loss": 2.3727,
"step": 372
},
{
"epoch": 0.6278140122028193,
"grad_norm": 0.0764508917927742,
"learning_rate": 1e-05,
"loss": 2.1992,
"step": 373
},
{
"epoch": 0.6294971596886177,
"grad_norm": 0.0799420177936554,
"learning_rate": 1e-05,
"loss": 2.2693,
"step": 374
},
{
"epoch": 0.6311803071744162,
"grad_norm": 0.06878554075956345,
"learning_rate": 1e-05,
"loss": 2.4749,
"step": 375
},
{
"epoch": 0.6328634546602147,
"grad_norm": 0.07085944712162018,
"learning_rate": 1e-05,
"loss": 2.3435,
"step": 376
},
{
"epoch": 0.634546602146013,
"grad_norm": 0.06489285826683044,
"learning_rate": 1e-05,
"loss": 2.3257,
"step": 377
},
{
"epoch": 0.6362297496318114,
"grad_norm": 0.06664973497390747,
"learning_rate": 1e-05,
"loss": 2.5022,
"step": 378
},
{
"epoch": 0.6379128971176099,
"grad_norm": 0.07660377770662308,
"learning_rate": 1e-05,
"loss": 2.3269,
"step": 379
},
{
"epoch": 0.6395960446034084,
"grad_norm": 0.06934674084186554,
"learning_rate": 1e-05,
"loss": 2.4021,
"step": 380
},
{
"epoch": 0.6412791920892068,
"grad_norm": 0.07515530288219452,
"learning_rate": 1e-05,
"loss": 2.3157,
"step": 381
},
{
"epoch": 0.6429623395750053,
"grad_norm": 0.07302498072385788,
"learning_rate": 1e-05,
"loss": 2.3892,
"step": 382
},
{
"epoch": 0.6446454870608037,
"grad_norm": 0.07303425669670105,
"learning_rate": 1e-05,
"loss": 2.3765,
"step": 383
},
{
"epoch": 0.6463286345466022,
"grad_norm": 0.07705460488796234,
"learning_rate": 1e-05,
"loss": 2.2684,
"step": 384
},
{
"epoch": 0.6480117820324006,
"grad_norm": 0.07487067580223083,
"learning_rate": 1e-05,
"loss": 2.3733,
"step": 385
},
{
"epoch": 0.6496949295181991,
"grad_norm": 0.06538619101047516,
"learning_rate": 1e-05,
"loss": 2.3789,
"step": 386
},
{
"epoch": 0.6513780770039975,
"grad_norm": 0.07406684756278992,
"learning_rate": 1e-05,
"loss": 2.332,
"step": 387
},
{
"epoch": 0.6530612244897959,
"grad_norm": 0.07246539741754532,
"learning_rate": 1e-05,
"loss": 2.2302,
"step": 388
},
{
"epoch": 0.6547443719755943,
"grad_norm": 0.07304323464632034,
"learning_rate": 1e-05,
"loss": 2.3708,
"step": 389
},
{
"epoch": 0.6564275194613928,
"grad_norm": 0.07457181811332703,
"learning_rate": 1e-05,
"loss": 2.2991,
"step": 390
},
{
"epoch": 0.6581106669471912,
"grad_norm": 0.07300930470228195,
"learning_rate": 1e-05,
"loss": 2.2423,
"step": 391
},
{
"epoch": 0.6597938144329897,
"grad_norm": 0.07508236914873123,
"learning_rate": 1e-05,
"loss": 2.2642,
"step": 392
},
{
"epoch": 0.6614769619187881,
"grad_norm": 0.07481173425912857,
"learning_rate": 1e-05,
"loss": 2.3,
"step": 393
},
{
"epoch": 0.6631601094045866,
"grad_norm": 0.06851742416620255,
"learning_rate": 1e-05,
"loss": 2.4534,
"step": 394
},
{
"epoch": 0.664843256890385,
"grad_norm": 0.07536716759204865,
"learning_rate": 1e-05,
"loss": 2.3264,
"step": 395
},
{
"epoch": 0.6665264043761835,
"grad_norm": 0.07752048969268799,
"learning_rate": 1e-05,
"loss": 2.4158,
"step": 396
},
{
"epoch": 0.6682095518619819,
"grad_norm": 0.06357281655073166,
"learning_rate": 1e-05,
"loss": 2.4956,
"step": 397
},
{
"epoch": 0.6698926993477804,
"grad_norm": 0.08333004266023636,
"learning_rate": 1e-05,
"loss": 2.3921,
"step": 398
},
{
"epoch": 0.6715758468335787,
"grad_norm": 0.06873282790184021,
"learning_rate": 1e-05,
"loss": 2.3611,
"step": 399
},
{
"epoch": 0.6732589943193772,
"grad_norm": 0.07533644139766693,
"learning_rate": 1e-05,
"loss": 2.3708,
"step": 400
},
{
"epoch": 0.6749421418051756,
"grad_norm": 0.07756076753139496,
"learning_rate": 1e-05,
"loss": 2.3003,
"step": 401
},
{
"epoch": 0.6766252892909741,
"grad_norm": 0.06644177436828613,
"learning_rate": 1e-05,
"loss": 2.4331,
"step": 402
},
{
"epoch": 0.6783084367767725,
"grad_norm": 0.07512148469686508,
"learning_rate": 1e-05,
"loss": 2.2881,
"step": 403
},
{
"epoch": 0.679991584262571,
"grad_norm": 0.08939874172210693,
"learning_rate": 1e-05,
"loss": 2.1564,
"step": 404
},
{
"epoch": 0.6816747317483695,
"grad_norm": 0.07984601706266403,
"learning_rate": 1e-05,
"loss": 2.3967,
"step": 405
},
{
"epoch": 0.6833578792341679,
"grad_norm": 0.0724392980337143,
"learning_rate": 1e-05,
"loss": 2.2859,
"step": 406
},
{
"epoch": 0.6850410267199664,
"grad_norm": 0.07025589793920517,
"learning_rate": 1e-05,
"loss": 2.3027,
"step": 407
},
{
"epoch": 0.6867241742057648,
"grad_norm": 0.07863828539848328,
"learning_rate": 1e-05,
"loss": 2.3286,
"step": 408
},
{
"epoch": 0.6884073216915633,
"grad_norm": 0.07466793060302734,
"learning_rate": 1e-05,
"loss": 2.2849,
"step": 409
},
{
"epoch": 0.6900904691773617,
"grad_norm": 0.07291209697723389,
"learning_rate": 1e-05,
"loss": 2.3931,
"step": 410
},
{
"epoch": 0.6917736166631601,
"grad_norm": 0.072298564016819,
"learning_rate": 1e-05,
"loss": 2.377,
"step": 411
},
{
"epoch": 0.6934567641489585,
"grad_norm": 0.06996294856071472,
"learning_rate": 1e-05,
"loss": 2.3503,
"step": 412
},
{
"epoch": 0.695139911634757,
"grad_norm": 0.07319701462984085,
"learning_rate": 1e-05,
"loss": 2.345,
"step": 413
},
{
"epoch": 0.6968230591205554,
"grad_norm": 0.0768033117055893,
"learning_rate": 1e-05,
"loss": 2.3679,
"step": 414
},
{
"epoch": 0.6985062066063539,
"grad_norm": 0.07401002943515778,
"learning_rate": 1e-05,
"loss": 2.3435,
"step": 415
},
{
"epoch": 0.7001893540921523,
"grad_norm": 0.07700485736131668,
"learning_rate": 1e-05,
"loss": 2.3428,
"step": 416
},
{
"epoch": 0.7018725015779508,
"grad_norm": 0.07446201890707016,
"learning_rate": 1e-05,
"loss": 2.4133,
"step": 417
},
{
"epoch": 0.7035556490637492,
"grad_norm": 0.06801878660917282,
"learning_rate": 1e-05,
"loss": 2.3665,
"step": 418
},
{
"epoch": 0.7052387965495477,
"grad_norm": 0.07989214360713959,
"learning_rate": 1e-05,
"loss": 2.3303,
"step": 419
},
{
"epoch": 0.7069219440353461,
"grad_norm": 0.07385462522506714,
"learning_rate": 1e-05,
"loss": 2.3608,
"step": 420
},
{
"epoch": 0.7086050915211446,
"grad_norm": 0.06808451563119888,
"learning_rate": 1e-05,
"loss": 2.4851,
"step": 421
},
{
"epoch": 0.7102882390069429,
"grad_norm": 0.07354162633419037,
"learning_rate": 1e-05,
"loss": 2.3005,
"step": 422
},
{
"epoch": 0.7119713864927414,
"grad_norm": 0.07730504870414734,
"learning_rate": 1e-05,
"loss": 2.2815,
"step": 423
},
{
"epoch": 0.7136545339785398,
"grad_norm": 0.08045239001512527,
"learning_rate": 1e-05,
"loss": 2.2695,
"step": 424
},
{
"epoch": 0.7153376814643383,
"grad_norm": 0.07997512817382812,
"learning_rate": 1e-05,
"loss": 2.3608,
"step": 425
},
{
"epoch": 0.7170208289501367,
"grad_norm": 0.07076172530651093,
"learning_rate": 1e-05,
"loss": 2.3411,
"step": 426
},
{
"epoch": 0.7187039764359352,
"grad_norm": 0.07223929464817047,
"learning_rate": 1e-05,
"loss": 2.3452,
"step": 427
},
{
"epoch": 0.7203871239217337,
"grad_norm": 0.07667456567287445,
"learning_rate": 1e-05,
"loss": 2.333,
"step": 428
},
{
"epoch": 0.7220702714075321,
"grad_norm": 0.07509643584489822,
"learning_rate": 1e-05,
"loss": 2.3701,
"step": 429
},
{
"epoch": 0.7237534188933306,
"grad_norm": 0.08230644464492798,
"learning_rate": 1e-05,
"loss": 2.3577,
"step": 430
},
{
"epoch": 0.725436566379129,
"grad_norm": 0.06938886642456055,
"learning_rate": 1e-05,
"loss": 2.4573,
"step": 431
},
{
"epoch": 0.7271197138649275,
"grad_norm": 0.07415178418159485,
"learning_rate": 1e-05,
"loss": 2.2834,
"step": 432
},
{
"epoch": 0.7288028613507258,
"grad_norm": 0.0821278989315033,
"learning_rate": 1e-05,
"loss": 2.2744,
"step": 433
},
{
"epoch": 0.7304860088365243,
"grad_norm": 0.07293502986431122,
"learning_rate": 1e-05,
"loss": 2.313,
"step": 434
},
{
"epoch": 0.7321691563223227,
"grad_norm": 0.07829819619655609,
"learning_rate": 1e-05,
"loss": 2.3849,
"step": 435
},
{
"epoch": 0.7338523038081212,
"grad_norm": 0.07795297354459763,
"learning_rate": 1e-05,
"loss": 2.2466,
"step": 436
},
{
"epoch": 0.7355354512939196,
"grad_norm": 0.06956803798675537,
"learning_rate": 1e-05,
"loss": 2.4038,
"step": 437
},
{
"epoch": 0.7372185987797181,
"grad_norm": 0.07948347926139832,
"learning_rate": 1e-05,
"loss": 2.3042,
"step": 438
},
{
"epoch": 0.7389017462655165,
"grad_norm": 0.08074218034744263,
"learning_rate": 1e-05,
"loss": 2.3314,
"step": 439
},
{
"epoch": 0.740584893751315,
"grad_norm": 0.08029188960790634,
"learning_rate": 1e-05,
"loss": 2.312,
"step": 440
},
{
"epoch": 0.7422680412371134,
"grad_norm": 0.0783049538731575,
"learning_rate": 1e-05,
"loss": 2.307,
"step": 441
},
{
"epoch": 0.7439511887229119,
"grad_norm": 0.08203115314245224,
"learning_rate": 1e-05,
"loss": 2.3081,
"step": 442
},
{
"epoch": 0.7456343362087103,
"grad_norm": 0.08666986972093582,
"learning_rate": 1e-05,
"loss": 2.3721,
"step": 443
},
{
"epoch": 0.7473174836945087,
"grad_norm": 0.08097022771835327,
"learning_rate": 1e-05,
"loss": 2.1912,
"step": 444
},
{
"epoch": 0.7490006311803071,
"grad_norm": 0.08272138237953186,
"learning_rate": 1e-05,
"loss": 2.3562,
"step": 445
},
{
"epoch": 0.7506837786661056,
"grad_norm": 0.08114828914403915,
"learning_rate": 1e-05,
"loss": 2.3569,
"step": 446
},
{
"epoch": 0.752366926151904,
"grad_norm": 0.07786712795495987,
"learning_rate": 1e-05,
"loss": 2.3772,
"step": 447
},
{
"epoch": 0.7540500736377025,
"grad_norm": 0.07603191584348679,
"learning_rate": 1e-05,
"loss": 2.2748,
"step": 448
},
{
"epoch": 0.7557332211235009,
"grad_norm": 0.08364319056272507,
"learning_rate": 1e-05,
"loss": 2.334,
"step": 449
},
{
"epoch": 0.7574163686092994,
"grad_norm": 0.07968125492334366,
"learning_rate": 1e-05,
"loss": 2.3225,
"step": 450
},
{
"epoch": 0.7590995160950978,
"grad_norm": 0.08204993605613708,
"learning_rate": 1e-05,
"loss": 2.3107,
"step": 451
},
{
"epoch": 0.7607826635808963,
"grad_norm": 0.08319111168384552,
"learning_rate": 1e-05,
"loss": 2.3994,
"step": 452
},
{
"epoch": 0.7624658110666948,
"grad_norm": 0.07812530547380447,
"learning_rate": 1e-05,
"loss": 2.2771,
"step": 453
},
{
"epoch": 0.7641489585524932,
"grad_norm": 0.07962696999311447,
"learning_rate": 1e-05,
"loss": 2.3094,
"step": 454
},
{
"epoch": 0.7658321060382917,
"grad_norm": 0.0815802663564682,
"learning_rate": 1e-05,
"loss": 2.3169,
"step": 455
},
{
"epoch": 0.76751525352409,
"grad_norm": 0.08460783958435059,
"learning_rate": 1e-05,
"loss": 2.2443,
"step": 456
},
{
"epoch": 0.7691984010098885,
"grad_norm": 0.07976390421390533,
"learning_rate": 1e-05,
"loss": 2.26,
"step": 457
},
{
"epoch": 0.7708815484956869,
"grad_norm": 0.08143635839223862,
"learning_rate": 1e-05,
"loss": 2.2517,
"step": 458
},
{
"epoch": 0.7725646959814854,
"grad_norm": 0.08004558831453323,
"learning_rate": 1e-05,
"loss": 2.3276,
"step": 459
},
{
"epoch": 0.7742478434672838,
"grad_norm": 0.0831751599907875,
"learning_rate": 1e-05,
"loss": 2.2842,
"step": 460
},
{
"epoch": 0.7759309909530823,
"grad_norm": 0.07613930851221085,
"learning_rate": 1e-05,
"loss": 2.3958,
"step": 461
},
{
"epoch": 0.7776141384388807,
"grad_norm": 0.08161590993404388,
"learning_rate": 1e-05,
"loss": 2.3287,
"step": 462
},
{
"epoch": 0.7792972859246792,
"grad_norm": 0.08616164326667786,
"learning_rate": 1e-05,
"loss": 2.3098,
"step": 463
},
{
"epoch": 0.7809804334104776,
"grad_norm": 0.08720822632312775,
"learning_rate": 1e-05,
"loss": 2.1388,
"step": 464
},
{
"epoch": 0.7826635808962761,
"grad_norm": 0.08598899841308594,
"learning_rate": 1e-05,
"loss": 2.3005,
"step": 465
},
{
"epoch": 0.7843467283820745,
"grad_norm": 0.07982167601585388,
"learning_rate": 1e-05,
"loss": 2.3049,
"step": 466
},
{
"epoch": 0.7860298758678729,
"grad_norm": 0.08733374625444412,
"learning_rate": 1e-05,
"loss": 2.2747,
"step": 467
},
{
"epoch": 0.7877130233536713,
"grad_norm": 0.08848235011100769,
"learning_rate": 1e-05,
"loss": 2.4331,
"step": 468
},
{
"epoch": 0.7893961708394698,
"grad_norm": 0.08619164675474167,
"learning_rate": 1e-05,
"loss": 2.2881,
"step": 469
},
{
"epoch": 0.7910793183252682,
"grad_norm": 0.08046075701713562,
"learning_rate": 1e-05,
"loss": 2.397,
"step": 470
},
{
"epoch": 0.7927624658110667,
"grad_norm": 0.08469874411821365,
"learning_rate": 1e-05,
"loss": 2.3225,
"step": 471
},
{
"epoch": 0.7944456132968651,
"grad_norm": 0.08878640830516815,
"learning_rate": 1e-05,
"loss": 2.2832,
"step": 472
},
{
"epoch": 0.7961287607826636,
"grad_norm": 0.08530005067586899,
"learning_rate": 1e-05,
"loss": 2.28,
"step": 473
},
{
"epoch": 0.797811908268462,
"grad_norm": 0.08089161664247513,
"learning_rate": 1e-05,
"loss": 2.2822,
"step": 474
},
{
"epoch": 0.7994950557542605,
"grad_norm": 0.0770372822880745,
"learning_rate": 1e-05,
"loss": 2.4031,
"step": 475
},
{
"epoch": 0.801178203240059,
"grad_norm": 0.08313820511102676,
"learning_rate": 1e-05,
"loss": 2.4009,
"step": 476
},
{
"epoch": 0.8028613507258574,
"grad_norm": 0.08684401214122772,
"learning_rate": 1e-05,
"loss": 2.4563,
"step": 477
},
{
"epoch": 0.8045444982116557,
"grad_norm": 0.08352997899055481,
"learning_rate": 1e-05,
"loss": 2.3242,
"step": 478
},
{
"epoch": 0.8062276456974542,
"grad_norm": 0.08148252218961716,
"learning_rate": 1e-05,
"loss": 2.3096,
"step": 479
},
{
"epoch": 0.8079107931832527,
"grad_norm": 0.08157838881015778,
"learning_rate": 1e-05,
"loss": 2.3108,
"step": 480
},
{
"epoch": 0.8095939406690511,
"grad_norm": 0.08561182022094727,
"learning_rate": 1e-05,
"loss": 2.2327,
"step": 481
},
{
"epoch": 0.8112770881548496,
"grad_norm": 0.09177689999341965,
"learning_rate": 1e-05,
"loss": 2.2129,
"step": 482
},
{
"epoch": 0.812960235640648,
"grad_norm": 0.08262176811695099,
"learning_rate": 1e-05,
"loss": 2.397,
"step": 483
},
{
"epoch": 0.8146433831264465,
"grad_norm": 0.08541447669267654,
"learning_rate": 1e-05,
"loss": 2.2419,
"step": 484
},
{
"epoch": 0.8163265306122449,
"grad_norm": 0.08732729405164719,
"learning_rate": 1e-05,
"loss": 2.3328,
"step": 485
},
{
"epoch": 0.8180096780980434,
"grad_norm": 0.08658833056688309,
"learning_rate": 1e-05,
"loss": 2.2793,
"step": 486
},
{
"epoch": 0.8196928255838418,
"grad_norm": 0.0789208933711052,
"learning_rate": 1e-05,
"loss": 2.4072,
"step": 487
},
{
"epoch": 0.8213759730696403,
"grad_norm": 0.07870952039957047,
"learning_rate": 1e-05,
"loss": 2.4082,
"step": 488
},
{
"epoch": 0.8230591205554386,
"grad_norm": 0.07583601027727127,
"learning_rate": 1e-05,
"loss": 2.3833,
"step": 489
},
{
"epoch": 0.8247422680412371,
"grad_norm": 0.08982661366462708,
"learning_rate": 1e-05,
"loss": 2.2766,
"step": 490
},
{
"epoch": 0.8264254155270355,
"grad_norm": 0.08841705322265625,
"learning_rate": 1e-05,
"loss": 2.2581,
"step": 491
},
{
"epoch": 0.828108563012834,
"grad_norm": 0.08784886449575424,
"learning_rate": 1e-05,
"loss": 2.2352,
"step": 492
},
{
"epoch": 0.8297917104986324,
"grad_norm": 0.08765432238578796,
"learning_rate": 1e-05,
"loss": 2.1957,
"step": 493
},
{
"epoch": 0.8314748579844309,
"grad_norm": 0.09070983529090881,
"learning_rate": 1e-05,
"loss": 2.2451,
"step": 494
},
{
"epoch": 0.8331580054702293,
"grad_norm": 0.08307146281003952,
"learning_rate": 1e-05,
"loss": 2.3645,
"step": 495
},
{
"epoch": 0.8348411529560278,
"grad_norm": 0.07774417847394943,
"learning_rate": 1e-05,
"loss": 2.3921,
"step": 496
},
{
"epoch": 0.8365243004418262,
"grad_norm": 0.08441779762506485,
"learning_rate": 1e-05,
"loss": 2.2974,
"step": 497
},
{
"epoch": 0.8382074479276247,
"grad_norm": 0.08773106336593628,
"learning_rate": 1e-05,
"loss": 2.3984,
"step": 498
},
{
"epoch": 0.8398905954134231,
"grad_norm": 0.08157604187726974,
"learning_rate": 1e-05,
"loss": 2.2946,
"step": 499
},
{
"epoch": 0.8415737428992216,
"grad_norm": 0.09280236810445786,
"learning_rate": 1e-05,
"loss": 2.3628,
"step": 500
},
{
"epoch": 0.8432568903850199,
"grad_norm": 0.08737549185752869,
"learning_rate": 1e-05,
"loss": 2.2593,
"step": 501
},
{
"epoch": 0.8449400378708184,
"grad_norm": 0.08917705714702606,
"learning_rate": 1e-05,
"loss": 2.2435,
"step": 502
},
{
"epoch": 0.8466231853566168,
"grad_norm": 0.08589258790016174,
"learning_rate": 1e-05,
"loss": 2.2869,
"step": 503
},
{
"epoch": 0.8483063328424153,
"grad_norm": 0.08363740891218185,
"learning_rate": 1e-05,
"loss": 2.1512,
"step": 504
},
{
"epoch": 0.8499894803282138,
"grad_norm": 0.09710842370986938,
"learning_rate": 1e-05,
"loss": 2.3042,
"step": 505
},
{
"epoch": 0.8516726278140122,
"grad_norm": 0.09031599014997482,
"learning_rate": 1e-05,
"loss": 2.2406,
"step": 506
},
{
"epoch": 0.8533557752998107,
"grad_norm": 0.08941849321126938,
"learning_rate": 1e-05,
"loss": 2.2725,
"step": 507
},
{
"epoch": 0.8550389227856091,
"grad_norm": 0.08926845341920853,
"learning_rate": 1e-05,
"loss": 2.323,
"step": 508
},
{
"epoch": 0.8567220702714076,
"grad_norm": 0.08846578001976013,
"learning_rate": 1e-05,
"loss": 2.3394,
"step": 509
},
{
"epoch": 0.858405217757206,
"grad_norm": 0.08452317863702774,
"learning_rate": 1e-05,
"loss": 2.4158,
"step": 510
},
{
"epoch": 0.8600883652430045,
"grad_norm": 0.08531490713357925,
"learning_rate": 1e-05,
"loss": 2.3113,
"step": 511
},
{
"epoch": 0.8617715127288028,
"grad_norm": 0.08221501857042313,
"learning_rate": 1e-05,
"loss": 2.3826,
"step": 512
},
{
"epoch": 0.8634546602146013,
"grad_norm": 0.08809410035610199,
"learning_rate": 1e-05,
"loss": 2.2666,
"step": 513
},
{
"epoch": 0.8651378077003997,
"grad_norm": 0.0881451964378357,
"learning_rate": 1e-05,
"loss": 2.4678,
"step": 514
},
{
"epoch": 0.8668209551861982,
"grad_norm": 0.0958879366517067,
"learning_rate": 1e-05,
"loss": 2.17,
"step": 515
},
{
"epoch": 0.8685041026719966,
"grad_norm": 0.08498766273260117,
"learning_rate": 1e-05,
"loss": 2.4021,
"step": 516
},
{
"epoch": 0.8701872501577951,
"grad_norm": 0.09182509779930115,
"learning_rate": 1e-05,
"loss": 2.2476,
"step": 517
},
{
"epoch": 0.8718703976435935,
"grad_norm": 0.08831535279750824,
"learning_rate": 1e-05,
"loss": 2.3013,
"step": 518
},
{
"epoch": 0.873553545129392,
"grad_norm": 0.08792266249656677,
"learning_rate": 1e-05,
"loss": 2.2463,
"step": 519
},
{
"epoch": 0.8752366926151904,
"grad_norm": 0.0804978460073471,
"learning_rate": 1e-05,
"loss": 2.5151,
"step": 520
},
{
"epoch": 0.8769198401009889,
"grad_norm": 0.09397967159748077,
"learning_rate": 1e-05,
"loss": 2.2487,
"step": 521
},
{
"epoch": 0.8786029875867873,
"grad_norm": 0.08882005512714386,
"learning_rate": 1e-05,
"loss": 2.225,
"step": 522
},
{
"epoch": 0.8802861350725857,
"grad_norm": 0.08365931361913681,
"learning_rate": 1e-05,
"loss": 2.4277,
"step": 523
},
{
"epoch": 0.8819692825583841,
"grad_norm": 0.08842651546001434,
"learning_rate": 1e-05,
"loss": 2.3884,
"step": 524
},
{
"epoch": 0.8836524300441826,
"grad_norm": 0.08760154247283936,
"learning_rate": 1e-05,
"loss": 2.2576,
"step": 525
},
{
"epoch": 0.885335577529981,
"grad_norm": 0.07843348383903503,
"learning_rate": 1e-05,
"loss": 2.4143,
"step": 526
},
{
"epoch": 0.8870187250157795,
"grad_norm": 0.09312726557254791,
"learning_rate": 1e-05,
"loss": 2.2472,
"step": 527
},
{
"epoch": 0.888701872501578,
"grad_norm": 0.09460542351007462,
"learning_rate": 1e-05,
"loss": 2.2043,
"step": 528
},
{
"epoch": 0.8903850199873764,
"grad_norm": 0.09200920909643173,
"learning_rate": 1e-05,
"loss": 2.3562,
"step": 529
},
{
"epoch": 0.8920681674731749,
"grad_norm": 0.08051000535488129,
"learning_rate": 1e-05,
"loss": 2.4146,
"step": 530
},
{
"epoch": 0.8937513149589733,
"grad_norm": 0.09969057142734528,
"learning_rate": 1e-05,
"loss": 2.3342,
"step": 531
},
{
"epoch": 0.8954344624447718,
"grad_norm": 0.08616895228624344,
"learning_rate": 1e-05,
"loss": 2.3381,
"step": 532
},
{
"epoch": 0.8971176099305702,
"grad_norm": 0.09115055203437805,
"learning_rate": 1e-05,
"loss": 2.2377,
"step": 533
},
{
"epoch": 0.8988007574163686,
"grad_norm": 0.10309138149023056,
"learning_rate": 1e-05,
"loss": 2.1418,
"step": 534
},
{
"epoch": 0.900483904902167,
"grad_norm": 0.09327155351638794,
"learning_rate": 1e-05,
"loss": 2.312,
"step": 535
},
{
"epoch": 0.9021670523879655,
"grad_norm": 0.09104789048433304,
"learning_rate": 1e-05,
"loss": 2.2759,
"step": 536
},
{
"epoch": 0.9038501998737639,
"grad_norm": 0.08858876675367355,
"learning_rate": 1e-05,
"loss": 2.4138,
"step": 537
},
{
"epoch": 0.9055333473595624,
"grad_norm": 0.08850864320993423,
"learning_rate": 1e-05,
"loss": 2.3915,
"step": 538
},
{
"epoch": 0.9072164948453608,
"grad_norm": 0.09071122854948044,
"learning_rate": 1e-05,
"loss": 2.4199,
"step": 539
},
{
"epoch": 0.9088996423311593,
"grad_norm": 0.08702193200588226,
"learning_rate": 1e-05,
"loss": 2.3079,
"step": 540
},
{
"epoch": 0.9105827898169577,
"grad_norm": 0.09564194083213806,
"learning_rate": 1e-05,
"loss": 2.2996,
"step": 541
},
{
"epoch": 0.9122659373027562,
"grad_norm": 0.08906988054513931,
"learning_rate": 1e-05,
"loss": 2.3958,
"step": 542
},
{
"epoch": 0.9139490847885546,
"grad_norm": 0.08117242157459259,
"learning_rate": 1e-05,
"loss": 2.5557,
"step": 543
},
{
"epoch": 0.9156322322743531,
"grad_norm": 0.09870729595422745,
"learning_rate": 1e-05,
"loss": 2.3542,
"step": 544
},
{
"epoch": 0.9173153797601514,
"grad_norm": 0.0906287208199501,
"learning_rate": 1e-05,
"loss": 2.2866,
"step": 545
},
{
"epoch": 0.9189985272459499,
"grad_norm": 0.08649491518735886,
"learning_rate": 1e-05,
"loss": 2.3547,
"step": 546
},
{
"epoch": 0.9206816747317483,
"grad_norm": 0.09572413563728333,
"learning_rate": 1e-05,
"loss": 2.377,
"step": 547
},
{
"epoch": 0.9223648222175468,
"grad_norm": 0.08862059563398361,
"learning_rate": 1e-05,
"loss": 2.3452,
"step": 548
},
{
"epoch": 0.9240479697033452,
"grad_norm": 0.09061957150697708,
"learning_rate": 1e-05,
"loss": 2.264,
"step": 549
},
{
"epoch": 0.9257311171891437,
"grad_norm": 0.10327678918838501,
"learning_rate": 1e-05,
"loss": 2.3362,
"step": 550
},
{
"epoch": 0.9274142646749421,
"grad_norm": 0.10101998597383499,
"learning_rate": 1e-05,
"loss": 2.2091,
"step": 551
},
{
"epoch": 0.9290974121607406,
"grad_norm": 0.08099676668643951,
"learning_rate": 1e-05,
"loss": 2.3779,
"step": 552
},
{
"epoch": 0.930780559646539,
"grad_norm": 0.09572342783212662,
"learning_rate": 1e-05,
"loss": 2.2186,
"step": 553
},
{
"epoch": 0.9324637071323375,
"grad_norm": 0.10440348833799362,
"learning_rate": 1e-05,
"loss": 2.2717,
"step": 554
},
{
"epoch": 0.934146854618136,
"grad_norm": 0.09859239310026169,
"learning_rate": 1e-05,
"loss": 2.2964,
"step": 555
},
{
"epoch": 0.9358300021039344,
"grad_norm": 0.08539914339780807,
"learning_rate": 1e-05,
"loss": 2.3541,
"step": 556
},
{
"epoch": 0.9375131495897328,
"grad_norm": 0.09667155891656876,
"learning_rate": 1e-05,
"loss": 2.2412,
"step": 557
},
{
"epoch": 0.9391962970755312,
"grad_norm": 0.09381328523159027,
"learning_rate": 1e-05,
"loss": 2.1632,
"step": 558
},
{
"epoch": 0.9408794445613297,
"grad_norm": 0.10293637216091156,
"learning_rate": 1e-05,
"loss": 2.2969,
"step": 559
},
{
"epoch": 0.9425625920471281,
"grad_norm": 0.08901844918727875,
"learning_rate": 1e-05,
"loss": 2.2806,
"step": 560
},
{
"epoch": 0.9442457395329266,
"grad_norm": 0.09931071847677231,
"learning_rate": 1e-05,
"loss": 2.2671,
"step": 561
},
{
"epoch": 0.945928887018725,
"grad_norm": 0.08619210124015808,
"learning_rate": 1e-05,
"loss": 2.428,
"step": 562
},
{
"epoch": 0.9476120345045235,
"grad_norm": 0.08460855484008789,
"learning_rate": 1e-05,
"loss": 2.2412,
"step": 563
},
{
"epoch": 0.9492951819903219,
"grad_norm": 0.09682973474264145,
"learning_rate": 1e-05,
"loss": 2.3339,
"step": 564
},
{
"epoch": 0.9509783294761204,
"grad_norm": 0.10189709812402725,
"learning_rate": 1e-05,
"loss": 2.2268,
"step": 565
},
{
"epoch": 0.9526614769619188,
"grad_norm": 0.10271991789340973,
"learning_rate": 1e-05,
"loss": 2.1819,
"step": 566
},
{
"epoch": 0.9543446244477173,
"grad_norm": 0.0901963859796524,
"learning_rate": 1e-05,
"loss": 2.3029,
"step": 567
},
{
"epoch": 0.9560277719335156,
"grad_norm": 0.09148905426263809,
"learning_rate": 1e-05,
"loss": 2.3362,
"step": 568
},
{
"epoch": 0.9577109194193141,
"grad_norm": 0.10434332489967346,
"learning_rate": 1e-05,
"loss": 2.3037,
"step": 569
},
{
"epoch": 0.9593940669051125,
"grad_norm": 0.0956675261259079,
"learning_rate": 1e-05,
"loss": 2.3442,
"step": 570
},
{
"epoch": 0.961077214390911,
"grad_norm": 0.09394146502017975,
"learning_rate": 1e-05,
"loss": 2.2913,
"step": 571
},
{
"epoch": 0.9627603618767094,
"grad_norm": 0.09179794043302536,
"learning_rate": 1e-05,
"loss": 2.21,
"step": 572
},
{
"epoch": 0.9644435093625079,
"grad_norm": 0.09866604208946228,
"learning_rate": 1e-05,
"loss": 2.2721,
"step": 573
},
{
"epoch": 0.9661266568483063,
"grad_norm": 0.10069537162780762,
"learning_rate": 1e-05,
"loss": 2.1637,
"step": 574
},
{
"epoch": 0.9678098043341048,
"grad_norm": 0.0923682376742363,
"learning_rate": 1e-05,
"loss": 2.2343,
"step": 575
},
{
"epoch": 0.9694929518199032,
"grad_norm": 0.08836492151021957,
"learning_rate": 1e-05,
"loss": 2.3794,
"step": 576
},
{
"epoch": 0.9711760993057017,
"grad_norm": 0.0894513726234436,
"learning_rate": 1e-05,
"loss": 2.2378,
"step": 577
},
{
"epoch": 0.9728592467915002,
"grad_norm": 0.08647426962852478,
"learning_rate": 1e-05,
"loss": 2.3589,
"step": 578
},
{
"epoch": 0.9745423942772985,
"grad_norm": 0.11035202443599701,
"learning_rate": 1e-05,
"loss": 2.2371,
"step": 579
},
{
"epoch": 0.976225541763097,
"grad_norm": 0.09551876783370972,
"learning_rate": 1e-05,
"loss": 2.3353,
"step": 580
},
{
"epoch": 0.9779086892488954,
"grad_norm": 0.0911082923412323,
"learning_rate": 1e-05,
"loss": 2.3264,
"step": 581
},
{
"epoch": 0.9795918367346939,
"grad_norm": 0.10280529409646988,
"learning_rate": 1e-05,
"loss": 2.2351,
"step": 582
},
{
"epoch": 0.9812749842204923,
"grad_norm": 0.09424940496683121,
"learning_rate": 1e-05,
"loss": 2.3464,
"step": 583
},
{
"epoch": 0.9829581317062908,
"grad_norm": 0.092115618288517,
"learning_rate": 1e-05,
"loss": 2.2799,
"step": 584
},
{
"epoch": 0.9846412791920892,
"grad_norm": 0.09771659225225449,
"learning_rate": 1e-05,
"loss": 2.3777,
"step": 585
},
{
"epoch": 0.9863244266778877,
"grad_norm": 0.09877105802297592,
"learning_rate": 1e-05,
"loss": 2.3613,
"step": 586
},
{
"epoch": 0.9880075741636861,
"grad_norm": 0.09816967695951462,
"learning_rate": 1e-05,
"loss": 2.2925,
"step": 587
},
{
"epoch": 0.9896907216494846,
"grad_norm": 0.0874725803732872,
"learning_rate": 1e-05,
"loss": 2.3154,
"step": 588
},
{
"epoch": 0.991373869135283,
"grad_norm": 0.09336823225021362,
"learning_rate": 1e-05,
"loss": 2.3933,
"step": 589
},
{
"epoch": 0.9930570166210814,
"grad_norm": 0.10439187288284302,
"learning_rate": 1e-05,
"loss": 2.3655,
"step": 590
},
{
"epoch": 0.9947401641068798,
"grad_norm": 0.09005751460790634,
"learning_rate": 1e-05,
"loss": 2.2971,
"step": 591
},
{
"epoch": 0.9964233115926783,
"grad_norm": 0.10612068325281143,
"learning_rate": 1e-05,
"loss": 2.3584,
"step": 592
},
{
"epoch": 0.9981064590784767,
"grad_norm": 0.09101177752017975,
"learning_rate": 1e-05,
"loss": 2.4402,
"step": 593
},
{
"epoch": 0.9997896065642752,
"grad_norm": 0.09874800592660904,
"learning_rate": 1e-05,
"loss": 2.326,
"step": 594
},
{
"epoch": 1.0014727540500736,
"grad_norm": 0.1025647521018982,
"learning_rate": 1e-05,
"loss": 2.4041,
"step": 595
},
{
"epoch": 1.003155901535872,
"grad_norm": 0.11109832674264908,
"learning_rate": 1e-05,
"loss": 2.2881,
"step": 596
},
{
"epoch": 1.0048390490216705,
"grad_norm": 0.09670565277338028,
"learning_rate": 1e-05,
"loss": 2.2003,
"step": 597
},
{
"epoch": 1.0065221965074689,
"grad_norm": 0.09513822942972183,
"learning_rate": 1e-05,
"loss": 2.3225,
"step": 598
},
{
"epoch": 1.0082053439932674,
"grad_norm": 0.11121483892202377,
"learning_rate": 1e-05,
"loss": 2.4143,
"step": 599
},
{
"epoch": 1.0098884914790658,
"grad_norm": 0.09941378980875015,
"learning_rate": 1e-05,
"loss": 2.333,
"step": 600
},
{
"epoch": 1.0115716389648644,
"grad_norm": 0.09730757772922516,
"learning_rate": 1e-05,
"loss": 2.3638,
"step": 601
},
{
"epoch": 1.0132547864506627,
"grad_norm": 0.10626422613859177,
"learning_rate": 1e-05,
"loss": 2.2303,
"step": 602
},
{
"epoch": 1.0149379339364613,
"grad_norm": 0.0958971306681633,
"learning_rate": 1e-05,
"loss": 2.3906,
"step": 603
},
{
"epoch": 1.0166210814222596,
"grad_norm": 0.10065159201622009,
"learning_rate": 1e-05,
"loss": 2.3425,
"step": 604
},
{
"epoch": 1.0183042289080582,
"grad_norm": 0.08671624213457108,
"learning_rate": 1e-05,
"loss": 2.2742,
"step": 605
},
{
"epoch": 1.0199873763938565,
"grad_norm": 0.09528376907110214,
"learning_rate": 1e-05,
"loss": 2.3765,
"step": 606
},
{
"epoch": 1.0216705238796548,
"grad_norm": 0.09153752028942108,
"learning_rate": 1e-05,
"loss": 2.2983,
"step": 607
},
{
"epoch": 1.0233536713654534,
"grad_norm": 0.10145740956068039,
"learning_rate": 1e-05,
"loss": 2.1774,
"step": 608
},
{
"epoch": 1.0250368188512518,
"grad_norm": 0.09908965229988098,
"learning_rate": 1e-05,
"loss": 2.3479,
"step": 609
},
{
"epoch": 1.0267199663370503,
"grad_norm": 0.09253786504268646,
"learning_rate": 1e-05,
"loss": 2.3228,
"step": 610
},
{
"epoch": 1.0284031138228487,
"grad_norm": 0.094690702855587,
"learning_rate": 1e-05,
"loss": 2.2864,
"step": 611
},
{
"epoch": 1.0300862613086472,
"grad_norm": 0.09160283952951431,
"learning_rate": 1e-05,
"loss": 2.4285,
"step": 612
},
{
"epoch": 1.0317694087944456,
"grad_norm": 0.10157333314418793,
"learning_rate": 1e-05,
"loss": 2.1316,
"step": 613
},
{
"epoch": 1.0334525562802441,
"grad_norm": 0.10498999804258347,
"learning_rate": 1e-05,
"loss": 2.373,
"step": 614
},
{
"epoch": 1.0351357037660425,
"grad_norm": 0.09599211066961288,
"learning_rate": 1e-05,
"loss": 2.3511,
"step": 615
},
{
"epoch": 1.036818851251841,
"grad_norm": 0.1121436059474945,
"learning_rate": 1e-05,
"loss": 2.127,
"step": 616
},
{
"epoch": 1.0385019987376394,
"grad_norm": 0.10269173234701157,
"learning_rate": 1e-05,
"loss": 2.2659,
"step": 617
},
{
"epoch": 1.040185146223438,
"grad_norm": 0.0945139229297638,
"learning_rate": 1e-05,
"loss": 2.3281,
"step": 618
},
{
"epoch": 1.0418682937092363,
"grad_norm": 0.09318878501653671,
"learning_rate": 1e-05,
"loss": 2.3247,
"step": 619
},
{
"epoch": 1.0435514411950346,
"grad_norm": 0.10471779108047485,
"learning_rate": 1e-05,
"loss": 2.3098,
"step": 620
},
{
"epoch": 1.0452345886808332,
"grad_norm": 0.10514305531978607,
"learning_rate": 1e-05,
"loss": 2.3647,
"step": 621
},
{
"epoch": 1.0469177361666315,
"grad_norm": 0.09875541925430298,
"learning_rate": 1e-05,
"loss": 2.4204,
"step": 622
},
{
"epoch": 1.04860088365243,
"grad_norm": 0.10112539678812027,
"learning_rate": 1e-05,
"loss": 2.3269,
"step": 623
},
{
"epoch": 1.0502840311382284,
"grad_norm": 0.09719318896532059,
"learning_rate": 1e-05,
"loss": 2.3223,
"step": 624
},
{
"epoch": 1.051967178624027,
"grad_norm": 0.09615301340818405,
"learning_rate": 1e-05,
"loss": 2.2798,
"step": 625
},
{
"epoch": 1.0536503261098253,
"grad_norm": 0.09600812941789627,
"learning_rate": 1e-05,
"loss": 2.2738,
"step": 626
},
{
"epoch": 1.055333473595624,
"grad_norm": 0.09326303005218506,
"learning_rate": 1e-05,
"loss": 2.23,
"step": 627
},
{
"epoch": 1.0570166210814222,
"grad_norm": 0.09689430892467499,
"learning_rate": 1e-05,
"loss": 2.2582,
"step": 628
},
{
"epoch": 1.0586997685672208,
"grad_norm": 0.10389314591884613,
"learning_rate": 1e-05,
"loss": 2.3733,
"step": 629
},
{
"epoch": 1.0603829160530192,
"grad_norm": 0.09320785105228424,
"learning_rate": 1e-05,
"loss": 2.3315,
"step": 630
},
{
"epoch": 1.0620660635388175,
"grad_norm": 0.10638166218996048,
"learning_rate": 1e-05,
"loss": 2.4058,
"step": 631
},
{
"epoch": 1.063749211024616,
"grad_norm": 0.09525519609451294,
"learning_rate": 1e-05,
"loss": 2.2803,
"step": 632
},
{
"epoch": 1.0654323585104144,
"grad_norm": 0.09904535114765167,
"learning_rate": 1e-05,
"loss": 2.3613,
"step": 633
},
{
"epoch": 1.067115505996213,
"grad_norm": 0.10914106667041779,
"learning_rate": 1e-05,
"loss": 2.3955,
"step": 634
},
{
"epoch": 1.0687986534820113,
"grad_norm": 0.10424593091011047,
"learning_rate": 1e-05,
"loss": 2.2332,
"step": 635
},
{
"epoch": 1.0704818009678099,
"grad_norm": 0.10360780358314514,
"learning_rate": 1e-05,
"loss": 2.3127,
"step": 636
},
{
"epoch": 1.0721649484536082,
"grad_norm": 0.11223631352186203,
"learning_rate": 1e-05,
"loss": 2.201,
"step": 637
},
{
"epoch": 1.0738480959394068,
"grad_norm": 0.09491337090730667,
"learning_rate": 1e-05,
"loss": 2.3129,
"step": 638
},
{
"epoch": 1.0755312434252051,
"grad_norm": 0.09244826436042786,
"learning_rate": 1e-05,
"loss": 2.3728,
"step": 639
},
{
"epoch": 1.0772143909110037,
"grad_norm": 0.0922231450676918,
"learning_rate": 1e-05,
"loss": 2.3225,
"step": 640
},
{
"epoch": 1.078897538396802,
"grad_norm": 0.10818596929311752,
"learning_rate": 1e-05,
"loss": 2.3104,
"step": 641
},
{
"epoch": 1.0805806858826004,
"grad_norm": 0.09497258812189102,
"learning_rate": 1e-05,
"loss": 2.3176,
"step": 642
},
{
"epoch": 1.082263833368399,
"grad_norm": 0.10034379363059998,
"learning_rate": 1e-05,
"loss": 2.3943,
"step": 643
},
{
"epoch": 1.0839469808541973,
"grad_norm": 0.10024038702249527,
"learning_rate": 1e-05,
"loss": 2.3127,
"step": 644
},
{
"epoch": 1.0856301283399958,
"grad_norm": 0.10074039548635483,
"learning_rate": 1e-05,
"loss": 2.2351,
"step": 645
},
{
"epoch": 1.0873132758257942,
"grad_norm": 0.09631813317537308,
"learning_rate": 1e-05,
"loss": 2.3101,
"step": 646
},
{
"epoch": 1.0889964233115927,
"grad_norm": 0.10632781684398651,
"learning_rate": 1e-05,
"loss": 2.3669,
"step": 647
},
{
"epoch": 1.090679570797391,
"grad_norm": 0.10795175284147263,
"learning_rate": 1e-05,
"loss": 2.3064,
"step": 648
},
{
"epoch": 1.0923627182831896,
"grad_norm": 0.11120691895484924,
"learning_rate": 1e-05,
"loss": 2.2911,
"step": 649
},
{
"epoch": 1.094045865768988,
"grad_norm": 0.10034749656915665,
"learning_rate": 1e-05,
"loss": 2.3696,
"step": 650
},
{
"epoch": 1.0957290132547866,
"grad_norm": 0.10955310612916946,
"learning_rate": 1e-05,
"loss": 2.3464,
"step": 651
},
{
"epoch": 1.097412160740585,
"grad_norm": 0.09739572554826736,
"learning_rate": 1e-05,
"loss": 2.325,
"step": 652
},
{
"epoch": 1.0990953082263832,
"grad_norm": 0.10152111947536469,
"learning_rate": 1e-05,
"loss": 2.3745,
"step": 653
},
{
"epoch": 1.1007784557121818,
"grad_norm": 0.10103686153888702,
"learning_rate": 1e-05,
"loss": 2.3303,
"step": 654
},
{
"epoch": 1.1024616031979801,
"grad_norm": 0.1003558412194252,
"learning_rate": 1e-05,
"loss": 2.312,
"step": 655
},
{
"epoch": 1.1041447506837787,
"grad_norm": 0.10518987476825714,
"learning_rate": 1e-05,
"loss": 2.3444,
"step": 656
},
{
"epoch": 1.105827898169577,
"grad_norm": 0.09896016865968704,
"learning_rate": 1e-05,
"loss": 2.2532,
"step": 657
},
{
"epoch": 1.1075110456553756,
"grad_norm": 0.09725090116262436,
"learning_rate": 1e-05,
"loss": 2.3625,
"step": 658
},
{
"epoch": 1.109194193141174,
"grad_norm": 0.09022284299135208,
"learning_rate": 1e-05,
"loss": 2.3743,
"step": 659
},
{
"epoch": 1.1108773406269725,
"grad_norm": 0.10471490770578384,
"learning_rate": 1e-05,
"loss": 2.3416,
"step": 660
},
{
"epoch": 1.1125604881127709,
"grad_norm": 0.10991263389587402,
"learning_rate": 1e-05,
"loss": 2.3214,
"step": 661
},
{
"epoch": 1.1142436355985694,
"grad_norm": 0.10231148451566696,
"learning_rate": 1e-05,
"loss": 2.2832,
"step": 662
},
{
"epoch": 1.1159267830843678,
"grad_norm": 0.09433937072753906,
"learning_rate": 1e-05,
"loss": 2.2645,
"step": 663
},
{
"epoch": 1.117609930570166,
"grad_norm": 0.13238666951656342,
"learning_rate": 1e-05,
"loss": 2.2483,
"step": 664
},
{
"epoch": 1.1192930780559647,
"grad_norm": 0.10956214368343353,
"learning_rate": 1e-05,
"loss": 2.2321,
"step": 665
},
{
"epoch": 1.120976225541763,
"grad_norm": 0.11065597832202911,
"learning_rate": 1e-05,
"loss": 2.1869,
"step": 666
},
{
"epoch": 1.1226593730275616,
"grad_norm": 0.10971678793430328,
"learning_rate": 1e-05,
"loss": 2.1855,
"step": 667
},
{
"epoch": 1.12434252051336,
"grad_norm": 0.11080143600702286,
"learning_rate": 1e-05,
"loss": 2.198,
"step": 668
},
{
"epoch": 1.1260256679991585,
"grad_norm": 0.10381001979112625,
"learning_rate": 1e-05,
"loss": 2.3384,
"step": 669
},
{
"epoch": 1.1277088154849568,
"grad_norm": 0.1026921421289444,
"learning_rate": 1e-05,
"loss": 2.2458,
"step": 670
},
{
"epoch": 1.1293919629707554,
"grad_norm": 0.10585295408964157,
"learning_rate": 1e-05,
"loss": 2.1859,
"step": 671
},
{
"epoch": 1.1310751104565537,
"grad_norm": 0.10650487244129181,
"learning_rate": 1e-05,
"loss": 2.2662,
"step": 672
},
{
"epoch": 1.1327582579423523,
"grad_norm": 0.10717649012804031,
"learning_rate": 1e-05,
"loss": 2.3088,
"step": 673
},
{
"epoch": 1.1344414054281506,
"grad_norm": 0.10479724407196045,
"learning_rate": 1e-05,
"loss": 2.3042,
"step": 674
},
{
"epoch": 1.136124552913949,
"grad_norm": 0.10629065334796906,
"learning_rate": 1e-05,
"loss": 2.3481,
"step": 675
},
{
"epoch": 1.1378077003997475,
"grad_norm": 0.10375174880027771,
"learning_rate": 1e-05,
"loss": 2.3845,
"step": 676
},
{
"epoch": 1.1394908478855459,
"grad_norm": 0.10122872143983841,
"learning_rate": 1e-05,
"loss": 2.335,
"step": 677
},
{
"epoch": 1.1411739953713445,
"grad_norm": 0.09846247732639313,
"learning_rate": 1e-05,
"loss": 2.4028,
"step": 678
},
{
"epoch": 1.1428571428571428,
"grad_norm": 0.11501342058181763,
"learning_rate": 1e-05,
"loss": 2.2419,
"step": 679
},
{
"epoch": 1.1445402903429414,
"grad_norm": 0.11248493194580078,
"learning_rate": 1e-05,
"loss": 2.1294,
"step": 680
},
{
"epoch": 1.1462234378287397,
"grad_norm": 0.1141652762889862,
"learning_rate": 1e-05,
"loss": 2.2842,
"step": 681
},
{
"epoch": 1.1479065853145383,
"grad_norm": 0.10232444107532501,
"learning_rate": 1e-05,
"loss": 2.1798,
"step": 682
},
{
"epoch": 1.1495897328003366,
"grad_norm": 0.10624698549509048,
"learning_rate": 1e-05,
"loss": 2.2474,
"step": 683
},
{
"epoch": 1.1512728802861352,
"grad_norm": 0.10583934187889099,
"learning_rate": 1e-05,
"loss": 2.2917,
"step": 684
},
{
"epoch": 1.1529560277719335,
"grad_norm": 0.10667344182729721,
"learning_rate": 1e-05,
"loss": 2.2581,
"step": 685
},
{
"epoch": 1.1546391752577319,
"grad_norm": 0.10415381193161011,
"learning_rate": 1e-05,
"loss": 2.3325,
"step": 686
},
{
"epoch": 1.1563223227435304,
"grad_norm": 0.109574094414711,
"learning_rate": 1e-05,
"loss": 2.3306,
"step": 687
},
{
"epoch": 1.1580054702293288,
"grad_norm": 0.10537154227495193,
"learning_rate": 1e-05,
"loss": 2.3396,
"step": 688
},
{
"epoch": 1.1596886177151273,
"grad_norm": 0.10670781880617142,
"learning_rate": 1e-05,
"loss": 2.2518,
"step": 689
},
{
"epoch": 1.1613717652009257,
"grad_norm": 0.10296822339296341,
"learning_rate": 1e-05,
"loss": 2.3911,
"step": 690
},
{
"epoch": 1.1630549126867242,
"grad_norm": 0.10323610156774521,
"learning_rate": 1e-05,
"loss": 2.415,
"step": 691
},
{
"epoch": 1.1647380601725226,
"grad_norm": 0.09952528029680252,
"learning_rate": 1e-05,
"loss": 2.3674,
"step": 692
},
{
"epoch": 1.1664212076583211,
"grad_norm": 0.10683920234441757,
"learning_rate": 1e-05,
"loss": 2.1606,
"step": 693
},
{
"epoch": 1.1681043551441195,
"grad_norm": 0.10594907402992249,
"learning_rate": 1e-05,
"loss": 2.3633,
"step": 694
},
{
"epoch": 1.169787502629918,
"grad_norm": 0.1164483055472374,
"learning_rate": 1e-05,
"loss": 2.272,
"step": 695
},
{
"epoch": 1.1714706501157164,
"grad_norm": 0.1053275316953659,
"learning_rate": 1e-05,
"loss": 2.3361,
"step": 696
},
{
"epoch": 1.1731537976015147,
"grad_norm": 0.11722961068153381,
"learning_rate": 1e-05,
"loss": 2.1008,
"step": 697
},
{
"epoch": 1.1748369450873133,
"grad_norm": 0.11388476192951202,
"learning_rate": 1e-05,
"loss": 2.3129,
"step": 698
},
{
"epoch": 1.1765200925731116,
"grad_norm": 0.1149948239326477,
"learning_rate": 1e-05,
"loss": 2.3503,
"step": 699
},
{
"epoch": 1.1782032400589102,
"grad_norm": 0.09305736422538757,
"learning_rate": 1e-05,
"loss": 2.3811,
"step": 700
},
{
"epoch": 1.1798863875447085,
"grad_norm": 0.1027708575129509,
"learning_rate": 1e-05,
"loss": 2.3262,
"step": 701
},
{
"epoch": 1.181569535030507,
"grad_norm": 0.1058826595544815,
"learning_rate": 1e-05,
"loss": 2.2576,
"step": 702
},
{
"epoch": 1.1832526825163054,
"grad_norm": 0.1003696396946907,
"learning_rate": 1e-05,
"loss": 2.2759,
"step": 703
},
{
"epoch": 1.184935830002104,
"grad_norm": 0.11113473027944565,
"learning_rate": 1e-05,
"loss": 2.4163,
"step": 704
},
{
"epoch": 1.1866189774879023,
"grad_norm": 0.10945228487253189,
"learning_rate": 1e-05,
"loss": 2.2725,
"step": 705
},
{
"epoch": 1.188302124973701,
"grad_norm": 0.1079326868057251,
"learning_rate": 1e-05,
"loss": 2.3048,
"step": 706
},
{
"epoch": 1.1899852724594993,
"grad_norm": 0.10752802342176437,
"learning_rate": 1e-05,
"loss": 2.2145,
"step": 707
},
{
"epoch": 1.1916684199452976,
"grad_norm": 0.10588284581899643,
"learning_rate": 1e-05,
"loss": 2.3025,
"step": 708
},
{
"epoch": 1.1933515674310962,
"grad_norm": 0.1051083654165268,
"learning_rate": 1e-05,
"loss": 2.3198,
"step": 709
},
{
"epoch": 1.1950347149168945,
"grad_norm": 0.11915988475084305,
"learning_rate": 1e-05,
"loss": 2.2456,
"step": 710
},
{
"epoch": 1.196717862402693,
"grad_norm": 0.10947719216346741,
"learning_rate": 1e-05,
"loss": 2.3479,
"step": 711
},
{
"epoch": 1.1984010098884914,
"grad_norm": 0.11522776633501053,
"learning_rate": 1e-05,
"loss": 2.2898,
"step": 712
},
{
"epoch": 1.20008415737429,
"grad_norm": 0.10741020739078522,
"learning_rate": 1e-05,
"loss": 2.3198,
"step": 713
},
{
"epoch": 1.2017673048600883,
"grad_norm": 0.10589215159416199,
"learning_rate": 1e-05,
"loss": 2.2812,
"step": 714
},
{
"epoch": 1.2034504523458869,
"grad_norm": 0.10151232033967972,
"learning_rate": 1e-05,
"loss": 2.429,
"step": 715
},
{
"epoch": 1.2051335998316852,
"grad_norm": 0.11951622366905212,
"learning_rate": 1e-05,
"loss": 2.1932,
"step": 716
},
{
"epoch": 1.2068167473174838,
"grad_norm": 0.11722715198993683,
"learning_rate": 1e-05,
"loss": 2.2356,
"step": 717
},
{
"epoch": 1.2084998948032821,
"grad_norm": 0.11441315710544586,
"learning_rate": 1e-05,
"loss": 2.2891,
"step": 718
},
{
"epoch": 1.2101830422890805,
"grad_norm": 0.10936987400054932,
"learning_rate": 1e-05,
"loss": 2.2843,
"step": 719
},
{
"epoch": 1.211866189774879,
"grad_norm": 0.12374020367860794,
"learning_rate": 1e-05,
"loss": 2.2944,
"step": 720
},
{
"epoch": 1.2135493372606774,
"grad_norm": 0.11024117469787598,
"learning_rate": 1e-05,
"loss": 2.2595,
"step": 721
},
{
"epoch": 1.215232484746476,
"grad_norm": 0.09707245975732803,
"learning_rate": 1e-05,
"loss": 2.3867,
"step": 722
},
{
"epoch": 1.2169156322322743,
"grad_norm": 0.11022404581308365,
"learning_rate": 1e-05,
"loss": 2.375,
"step": 723
},
{
"epoch": 1.2185987797180728,
"grad_norm": 0.10732002556324005,
"learning_rate": 1e-05,
"loss": 2.3674,
"step": 724
},
{
"epoch": 1.2202819272038712,
"grad_norm": 0.11548677086830139,
"learning_rate": 1e-05,
"loss": 2.3284,
"step": 725
},
{
"epoch": 1.2219650746896698,
"grad_norm": 0.10313412547111511,
"learning_rate": 1e-05,
"loss": 2.4128,
"step": 726
},
{
"epoch": 1.223648222175468,
"grad_norm": 0.12717945873737335,
"learning_rate": 1e-05,
"loss": 2.2847,
"step": 727
},
{
"epoch": 1.2253313696612667,
"grad_norm": 0.11565182358026505,
"learning_rate": 1e-05,
"loss": 2.2695,
"step": 728
},
{
"epoch": 1.227014517147065,
"grad_norm": 0.10489466041326523,
"learning_rate": 1e-05,
"loss": 2.3394,
"step": 729
},
{
"epoch": 1.2286976646328633,
"grad_norm": 0.11056289076805115,
"learning_rate": 1e-05,
"loss": 2.4165,
"step": 730
},
{
"epoch": 1.230380812118662,
"grad_norm": 0.12048956751823425,
"learning_rate": 1e-05,
"loss": 2.2289,
"step": 731
},
{
"epoch": 1.2320639596044602,
"grad_norm": 0.10263136774301529,
"learning_rate": 1e-05,
"loss": 2.3306,
"step": 732
},
{
"epoch": 1.2337471070902588,
"grad_norm": 0.11179950088262558,
"learning_rate": 1e-05,
"loss": 2.3481,
"step": 733
},
{
"epoch": 1.2354302545760572,
"grad_norm": 0.10484311729669571,
"learning_rate": 1e-05,
"loss": 2.2703,
"step": 734
},
{
"epoch": 1.2371134020618557,
"grad_norm": 0.1182483434677124,
"learning_rate": 1e-05,
"loss": 2.2328,
"step": 735
},
{
"epoch": 1.238796549547654,
"grad_norm": 0.11377429217100143,
"learning_rate": 1e-05,
"loss": 2.3657,
"step": 736
},
{
"epoch": 1.2404796970334526,
"grad_norm": 0.11151503771543503,
"learning_rate": 1e-05,
"loss": 2.3542,
"step": 737
},
{
"epoch": 1.242162844519251,
"grad_norm": 0.12628555297851562,
"learning_rate": 1e-05,
"loss": 2.2634,
"step": 738
},
{
"epoch": 1.2438459920050495,
"grad_norm": 0.10311713814735413,
"learning_rate": 1e-05,
"loss": 2.2717,
"step": 739
},
{
"epoch": 1.2455291394908479,
"grad_norm": 0.12768767774105072,
"learning_rate": 1e-05,
"loss": 2.1725,
"step": 740
},
{
"epoch": 1.2472122869766462,
"grad_norm": 0.12390502542257309,
"learning_rate": 1e-05,
"loss": 2.1708,
"step": 741
},
{
"epoch": 1.2488954344624448,
"grad_norm": 0.10566207021474838,
"learning_rate": 1e-05,
"loss": 2.3469,
"step": 742
},
{
"epoch": 1.2505785819482433,
"grad_norm": 0.10176009684801102,
"learning_rate": 1e-05,
"loss": 2.3159,
"step": 743
},
{
"epoch": 1.2522617294340417,
"grad_norm": 0.10881732404232025,
"learning_rate": 1e-05,
"loss": 2.2966,
"step": 744
},
{
"epoch": 1.25394487691984,
"grad_norm": 0.11917608976364136,
"learning_rate": 1e-05,
"loss": 2.395,
"step": 745
},
{
"epoch": 1.2556280244056386,
"grad_norm": 0.09600858390331268,
"learning_rate": 1e-05,
"loss": 2.3479,
"step": 746
},
{
"epoch": 1.257311171891437,
"grad_norm": 0.11550504714250565,
"learning_rate": 1e-05,
"loss": 2.301,
"step": 747
},
{
"epoch": 1.2589943193772355,
"grad_norm": 0.10588584840297699,
"learning_rate": 1e-05,
"loss": 2.4163,
"step": 748
},
{
"epoch": 1.2606774668630338,
"grad_norm": 0.10998673737049103,
"learning_rate": 1e-05,
"loss": 2.3379,
"step": 749
},
{
"epoch": 1.2623606143488324,
"grad_norm": 0.10513128340244293,
"learning_rate": 1e-05,
"loss": 2.3795,
"step": 750
},
{
"epoch": 1.2640437618346307,
"grad_norm": 0.11185754835605621,
"learning_rate": 1e-05,
"loss": 2.2583,
"step": 751
},
{
"epoch": 1.265726909320429,
"grad_norm": 0.10794227570295334,
"learning_rate": 1e-05,
"loss": 2.285,
"step": 752
},
{
"epoch": 1.2674100568062276,
"grad_norm": 0.12522459030151367,
"learning_rate": 1e-05,
"loss": 2.2292,
"step": 753
},
{
"epoch": 1.2690932042920262,
"grad_norm": 0.11628364026546478,
"learning_rate": 1e-05,
"loss": 2.3342,
"step": 754
},
{
"epoch": 1.2707763517778246,
"grad_norm": 0.12842795252799988,
"learning_rate": 1e-05,
"loss": 2.1455,
"step": 755
},
{
"epoch": 1.272459499263623,
"grad_norm": 0.11268262565135956,
"learning_rate": 1e-05,
"loss": 2.2241,
"step": 756
},
{
"epoch": 1.2741426467494215,
"grad_norm": 0.11674508452415466,
"learning_rate": 1e-05,
"loss": 2.2677,
"step": 757
},
{
"epoch": 1.2758257942352198,
"grad_norm": 0.11475373059511185,
"learning_rate": 1e-05,
"loss": 2.4075,
"step": 758
},
{
"epoch": 1.2775089417210184,
"grad_norm": 0.11378497630357742,
"learning_rate": 1e-05,
"loss": 2.3032,
"step": 759
},
{
"epoch": 1.2791920892068167,
"grad_norm": 0.10426255315542221,
"learning_rate": 1e-05,
"loss": 2.2488,
"step": 760
},
{
"epoch": 1.2808752366926153,
"grad_norm": 0.11820263415575027,
"learning_rate": 1e-05,
"loss": 2.197,
"step": 761
},
{
"epoch": 1.2825583841784136,
"grad_norm": 0.10741489380598068,
"learning_rate": 1e-05,
"loss": 2.2811,
"step": 762
},
{
"epoch": 1.284241531664212,
"grad_norm": 0.115534208714962,
"learning_rate": 1e-05,
"loss": 2.3105,
"step": 763
},
{
"epoch": 1.2859246791500105,
"grad_norm": 0.1159248948097229,
"learning_rate": 1e-05,
"loss": 2.2963,
"step": 764
},
{
"epoch": 1.287607826635809,
"grad_norm": 0.11940732598304749,
"learning_rate": 1e-05,
"loss": 2.3274,
"step": 765
},
{
"epoch": 1.2892909741216074,
"grad_norm": 0.11882008612155914,
"learning_rate": 1e-05,
"loss": 2.2405,
"step": 766
},
{
"epoch": 1.2909741216074058,
"grad_norm": 0.10939499735832214,
"learning_rate": 1e-05,
"loss": 2.3008,
"step": 767
},
{
"epoch": 1.2926572690932043,
"grad_norm": 0.11414020508527756,
"learning_rate": 1e-05,
"loss": 2.3164,
"step": 768
},
{
"epoch": 1.2943404165790027,
"grad_norm": 0.11446741968393326,
"learning_rate": 1e-05,
"loss": 2.2524,
"step": 769
},
{
"epoch": 1.2960235640648012,
"grad_norm": 0.12233757227659225,
"learning_rate": 1e-05,
"loss": 2.3997,
"step": 770
},
{
"epoch": 1.2977067115505996,
"grad_norm": 0.11746780574321747,
"learning_rate": 1e-05,
"loss": 2.2241,
"step": 771
},
{
"epoch": 1.2993898590363981,
"grad_norm": 0.12653754651546478,
"learning_rate": 1e-05,
"loss": 2.2181,
"step": 772
},
{
"epoch": 1.3010730065221965,
"grad_norm": 0.11092430353164673,
"learning_rate": 1e-05,
"loss": 2.194,
"step": 773
},
{
"epoch": 1.3027561540079948,
"grad_norm": 0.11273445188999176,
"learning_rate": 1e-05,
"loss": 2.2821,
"step": 774
},
{
"epoch": 1.3044393014937934,
"grad_norm": 0.10755831003189087,
"learning_rate": 1e-05,
"loss": 2.3381,
"step": 775
},
{
"epoch": 1.306122448979592,
"grad_norm": 0.10324183851480484,
"learning_rate": 1e-05,
"loss": 2.4531,
"step": 776
},
{
"epoch": 1.3078055964653903,
"grad_norm": 0.1238187626004219,
"learning_rate": 1e-05,
"loss": 2.2378,
"step": 777
},
{
"epoch": 1.3094887439511886,
"grad_norm": 0.10919329524040222,
"learning_rate": 1e-05,
"loss": 2.3157,
"step": 778
},
{
"epoch": 1.3111718914369872,
"grad_norm": 0.11661651730537415,
"learning_rate": 1e-05,
"loss": 2.3889,
"step": 779
},
{
"epoch": 1.3128550389227855,
"grad_norm": 0.11324804276227951,
"learning_rate": 1e-05,
"loss": 2.366,
"step": 780
},
{
"epoch": 1.314538186408584,
"grad_norm": 0.11539211124181747,
"learning_rate": 1e-05,
"loss": 2.2661,
"step": 781
},
{
"epoch": 1.3162213338943825,
"grad_norm": 0.12013803422451019,
"learning_rate": 1e-05,
"loss": 2.2388,
"step": 782
},
{
"epoch": 1.317904481380181,
"grad_norm": 0.1297876238822937,
"learning_rate": 1e-05,
"loss": 2.338,
"step": 783
},
{
"epoch": 1.3195876288659794,
"grad_norm": 0.11792443692684174,
"learning_rate": 1e-05,
"loss": 2.3162,
"step": 784
},
{
"epoch": 1.3212707763517777,
"grad_norm": 0.11543410271406174,
"learning_rate": 1e-05,
"loss": 2.325,
"step": 785
},
{
"epoch": 1.3229539238375763,
"grad_norm": 0.11507069319486618,
"learning_rate": 1e-05,
"loss": 2.3389,
"step": 786
},
{
"epoch": 1.3246370713233748,
"grad_norm": 0.11883421987295151,
"learning_rate": 1e-05,
"loss": 2.3784,
"step": 787
},
{
"epoch": 1.3263202188091732,
"grad_norm": 0.11997753381729126,
"learning_rate": 1e-05,
"loss": 2.2183,
"step": 788
},
{
"epoch": 1.3280033662949715,
"grad_norm": 0.12312667816877365,
"learning_rate": 1e-05,
"loss": 2.2661,
"step": 789
},
{
"epoch": 1.32968651378077,
"grad_norm": 0.1280994415283203,
"learning_rate": 1e-05,
"loss": 2.235,
"step": 790
},
{
"epoch": 1.3313696612665684,
"grad_norm": 0.12460897862911224,
"learning_rate": 1e-05,
"loss": 2.2775,
"step": 791
},
{
"epoch": 1.333052808752367,
"grad_norm": 0.11441405862569809,
"learning_rate": 1e-05,
"loss": 2.2642,
"step": 792
},
{
"epoch": 1.3347359562381653,
"grad_norm": 0.1078685000538826,
"learning_rate": 1e-05,
"loss": 2.3174,
"step": 793
},
{
"epoch": 1.3364191037239639,
"grad_norm": 0.11945922672748566,
"learning_rate": 1e-05,
"loss": 2.3101,
"step": 794
},
{
"epoch": 1.3381022512097622,
"grad_norm": 0.11506087332963943,
"learning_rate": 1e-05,
"loss": 2.3167,
"step": 795
},
{
"epoch": 1.3397853986955606,
"grad_norm": 0.12365138530731201,
"learning_rate": 1e-05,
"loss": 2.3044,
"step": 796
},
{
"epoch": 1.3414685461813591,
"grad_norm": 0.12331211566925049,
"learning_rate": 1e-05,
"loss": 2.2058,
"step": 797
},
{
"epoch": 1.3431516936671577,
"grad_norm": 0.12298640608787537,
"learning_rate": 1e-05,
"loss": 2.21,
"step": 798
},
{
"epoch": 1.344834841152956,
"grad_norm": 0.12047012150287628,
"learning_rate": 1e-05,
"loss": 2.2781,
"step": 799
},
{
"epoch": 1.3465179886387544,
"grad_norm": 0.12428031861782074,
"learning_rate": 1e-05,
"loss": 2.3032,
"step": 800
},
{
"epoch": 1.348201136124553,
"grad_norm": 0.1128249540925026,
"learning_rate": 1e-05,
"loss": 2.3135,
"step": 801
},
{
"epoch": 1.3498842836103513,
"grad_norm": 0.12616464495658875,
"learning_rate": 1e-05,
"loss": 2.1487,
"step": 802
},
{
"epoch": 1.3515674310961499,
"grad_norm": 0.11388704925775528,
"learning_rate": 1e-05,
"loss": 2.2346,
"step": 803
},
{
"epoch": 1.3532505785819482,
"grad_norm": 0.10213828831911087,
"learning_rate": 1e-05,
"loss": 2.2859,
"step": 804
},
{
"epoch": 1.3549337260677468,
"grad_norm": 0.1226121038198471,
"learning_rate": 1e-05,
"loss": 2.2183,
"step": 805
},
{
"epoch": 1.356616873553545,
"grad_norm": 0.11445735394954681,
"learning_rate": 1e-05,
"loss": 2.3784,
"step": 806
},
{
"epoch": 1.3583000210393434,
"grad_norm": 0.11648505181074142,
"learning_rate": 1e-05,
"loss": 2.3442,
"step": 807
},
{
"epoch": 1.359983168525142,
"grad_norm": 0.1296563744544983,
"learning_rate": 1e-05,
"loss": 2.2469,
"step": 808
},
{
"epoch": 1.3616663160109406,
"grad_norm": 0.12322400510311127,
"learning_rate": 1e-05,
"loss": 2.2915,
"step": 809
},
{
"epoch": 1.363349463496739,
"grad_norm": 0.11419309675693512,
"learning_rate": 1e-05,
"loss": 2.3024,
"step": 810
},
{
"epoch": 1.3650326109825373,
"grad_norm": 0.12253374606370926,
"learning_rate": 1e-05,
"loss": 2.2969,
"step": 811
},
{
"epoch": 1.3667157584683358,
"grad_norm": 0.1254422962665558,
"learning_rate": 1e-05,
"loss": 2.364,
"step": 812
},
{
"epoch": 1.3683989059541342,
"grad_norm": 0.12984994053840637,
"learning_rate": 1e-05,
"loss": 2.2936,
"step": 813
},
{
"epoch": 1.3700820534399327,
"grad_norm": 0.1182006224989891,
"learning_rate": 1e-05,
"loss": 2.2673,
"step": 814
},
{
"epoch": 1.371765200925731,
"grad_norm": 0.12920832633972168,
"learning_rate": 1e-05,
"loss": 2.1582,
"step": 815
},
{
"epoch": 1.3734483484115296,
"grad_norm": 0.1216689869761467,
"learning_rate": 1e-05,
"loss": 2.3479,
"step": 816
},
{
"epoch": 1.375131495897328,
"grad_norm": 0.12459319084882736,
"learning_rate": 1e-05,
"loss": 2.1868,
"step": 817
},
{
"epoch": 1.3768146433831263,
"grad_norm": 0.11144936084747314,
"learning_rate": 1e-05,
"loss": 2.3663,
"step": 818
},
{
"epoch": 1.3784977908689249,
"grad_norm": 0.1110294982790947,
"learning_rate": 1e-05,
"loss": 2.3164,
"step": 819
},
{
"epoch": 1.3801809383547234,
"grad_norm": 0.11903022974729538,
"learning_rate": 1e-05,
"loss": 2.2589,
"step": 820
},
{
"epoch": 1.3818640858405218,
"grad_norm": 0.10610275715589523,
"learning_rate": 1e-05,
"loss": 2.4153,
"step": 821
},
{
"epoch": 1.3835472333263201,
"grad_norm": 0.11972808837890625,
"learning_rate": 1e-05,
"loss": 2.3901,
"step": 822
},
{
"epoch": 1.3852303808121187,
"grad_norm": 0.10772975534200668,
"learning_rate": 1e-05,
"loss": 2.3555,
"step": 823
},
{
"epoch": 1.386913528297917,
"grad_norm": 0.11757270246744156,
"learning_rate": 1e-05,
"loss": 2.2677,
"step": 824
},
{
"epoch": 1.3885966757837156,
"grad_norm": 0.1217508539557457,
"learning_rate": 1e-05,
"loss": 2.2267,
"step": 825
},
{
"epoch": 1.390279823269514,
"grad_norm": 0.10996967554092407,
"learning_rate": 1e-05,
"loss": 2.3965,
"step": 826
},
{
"epoch": 1.3919629707553125,
"grad_norm": 0.13068005442619324,
"learning_rate": 1e-05,
"loss": 2.1991,
"step": 827
},
{
"epoch": 1.3936461182411108,
"grad_norm": 0.12149260193109512,
"learning_rate": 1e-05,
"loss": 2.2775,
"step": 828
},
{
"epoch": 1.3953292657269092,
"grad_norm": 0.1100870743393898,
"learning_rate": 1e-05,
"loss": 2.2571,
"step": 829
},
{
"epoch": 1.3970124132127077,
"grad_norm": 0.10005280375480652,
"learning_rate": 1e-05,
"loss": 2.2808,
"step": 830
},
{
"epoch": 1.3986955606985063,
"grad_norm": 0.11633820086717606,
"learning_rate": 1e-05,
"loss": 2.3215,
"step": 831
},
{
"epoch": 1.4003787081843047,
"grad_norm": 0.11901983618736267,
"learning_rate": 1e-05,
"loss": 2.4236,
"step": 832
},
{
"epoch": 1.402061855670103,
"grad_norm": 0.11173246055841446,
"learning_rate": 1e-05,
"loss": 2.3457,
"step": 833
},
{
"epoch": 1.4037450031559016,
"grad_norm": 0.10333243012428284,
"learning_rate": 1e-05,
"loss": 2.2659,
"step": 834
},
{
"epoch": 1.4054281506417,
"grad_norm": 0.13903972506523132,
"learning_rate": 1e-05,
"loss": 2.1946,
"step": 835
},
{
"epoch": 1.4071112981274985,
"grad_norm": 0.11832322925329208,
"learning_rate": 1e-05,
"loss": 2.3223,
"step": 836
},
{
"epoch": 1.4087944456132968,
"grad_norm": 0.10906493663787842,
"learning_rate": 1e-05,
"loss": 2.4316,
"step": 837
},
{
"epoch": 1.4104775930990954,
"grad_norm": 0.10980133712291718,
"learning_rate": 1e-05,
"loss": 2.3525,
"step": 838
},
{
"epoch": 1.4121607405848937,
"grad_norm": 0.12958386540412903,
"learning_rate": 1e-05,
"loss": 2.3081,
"step": 839
},
{
"epoch": 1.413843888070692,
"grad_norm": 0.1342059075832367,
"learning_rate": 1e-05,
"loss": 2.3564,
"step": 840
},
{
"epoch": 1.4155270355564906,
"grad_norm": 0.1362716406583786,
"learning_rate": 1e-05,
"loss": 2.2435,
"step": 841
},
{
"epoch": 1.4172101830422892,
"grad_norm": 0.10814797878265381,
"learning_rate": 1e-05,
"loss": 2.3373,
"step": 842
},
{
"epoch": 1.4188933305280875,
"grad_norm": 0.111182801425457,
"learning_rate": 1e-05,
"loss": 2.2921,
"step": 843
},
{
"epoch": 1.4205764780138859,
"grad_norm": 0.11161399632692337,
"learning_rate": 1e-05,
"loss": 2.3816,
"step": 844
},
{
"epoch": 1.4222596254996844,
"grad_norm": 0.1261526495218277,
"learning_rate": 1e-05,
"loss": 2.4082,
"step": 845
},
{
"epoch": 1.4239427729854828,
"grad_norm": 0.10805182158946991,
"learning_rate": 1e-05,
"loss": 2.3622,
"step": 846
},
{
"epoch": 1.4256259204712813,
"grad_norm": 0.12294517457485199,
"learning_rate": 1e-05,
"loss": 2.3638,
"step": 847
},
{
"epoch": 1.4273090679570797,
"grad_norm": 0.10903607308864594,
"learning_rate": 1e-05,
"loss": 2.3484,
"step": 848
},
{
"epoch": 1.4289922154428782,
"grad_norm": 0.12460491806268692,
"learning_rate": 1e-05,
"loss": 2.2046,
"step": 849
},
{
"epoch": 1.4306753629286766,
"grad_norm": 0.13793089985847473,
"learning_rate": 1e-05,
"loss": 2.2437,
"step": 850
},
{
"epoch": 1.4323585104144752,
"grad_norm": 0.11700379103422165,
"learning_rate": 1e-05,
"loss": 2.2288,
"step": 851
},
{
"epoch": 1.4340416579002735,
"grad_norm": 0.11343109607696533,
"learning_rate": 1e-05,
"loss": 2.2501,
"step": 852
},
{
"epoch": 1.435724805386072,
"grad_norm": 0.10918331891298294,
"learning_rate": 1e-05,
"loss": 2.47,
"step": 853
},
{
"epoch": 1.4374079528718704,
"grad_norm": 0.12782573699951172,
"learning_rate": 1e-05,
"loss": 2.2281,
"step": 854
},
{
"epoch": 1.4390911003576687,
"grad_norm": 0.12039442360401154,
"learning_rate": 1e-05,
"loss": 2.2766,
"step": 855
},
{
"epoch": 1.4407742478434673,
"grad_norm": 0.13949096202850342,
"learning_rate": 1e-05,
"loss": 2.198,
"step": 856
},
{
"epoch": 1.4424573953292656,
"grad_norm": 0.13327306509017944,
"learning_rate": 1e-05,
"loss": 2.2253,
"step": 857
},
{
"epoch": 1.4441405428150642,
"grad_norm": 0.1229238212108612,
"learning_rate": 1e-05,
"loss": 2.3147,
"step": 858
},
{
"epoch": 1.4458236903008626,
"grad_norm": 0.13407859206199646,
"learning_rate": 1e-05,
"loss": 2.2532,
"step": 859
},
{
"epoch": 1.4475068377866611,
"grad_norm": 0.1280384659767151,
"learning_rate": 1e-05,
"loss": 2.3174,
"step": 860
},
{
"epoch": 1.4491899852724595,
"grad_norm": 0.1532362997531891,
"learning_rate": 1e-05,
"loss": 2.1671,
"step": 861
},
{
"epoch": 1.450873132758258,
"grad_norm": 0.1134854182600975,
"learning_rate": 1e-05,
"loss": 2.3607,
"step": 862
},
{
"epoch": 1.4525562802440564,
"grad_norm": 0.11682198196649551,
"learning_rate": 1e-05,
"loss": 2.4041,
"step": 863
},
{
"epoch": 1.454239427729855,
"grad_norm": 0.11356412619352341,
"learning_rate": 1e-05,
"loss": 2.2756,
"step": 864
},
{
"epoch": 1.4559225752156533,
"grad_norm": 0.11278104037046432,
"learning_rate": 1e-05,
"loss": 2.2983,
"step": 865
},
{
"epoch": 1.4576057227014516,
"grad_norm": 0.13442599773406982,
"learning_rate": 1e-05,
"loss": 2.2593,
"step": 866
},
{
"epoch": 1.4592888701872502,
"grad_norm": 0.1254800707101822,
"learning_rate": 1e-05,
"loss": 2.3213,
"step": 867
},
{
"epoch": 1.4609720176730487,
"grad_norm": 0.12374315410852432,
"learning_rate": 1e-05,
"loss": 2.4221,
"step": 868
},
{
"epoch": 1.462655165158847,
"grad_norm": 0.13577024638652802,
"learning_rate": 1e-05,
"loss": 2.2473,
"step": 869
},
{
"epoch": 1.4643383126446454,
"grad_norm": 0.12822799384593964,
"learning_rate": 1e-05,
"loss": 2.3057,
"step": 870
},
{
"epoch": 1.466021460130444,
"grad_norm": 0.1283286213874817,
"learning_rate": 1e-05,
"loss": 2.374,
"step": 871
},
{
"epoch": 1.4677046076162423,
"grad_norm": 0.12054271996021271,
"learning_rate": 1e-05,
"loss": 2.3369,
"step": 872
},
{
"epoch": 1.469387755102041,
"grad_norm": 0.127189502120018,
"learning_rate": 1e-05,
"loss": 2.3167,
"step": 873
},
{
"epoch": 1.4710709025878392,
"grad_norm": 0.12767814099788666,
"learning_rate": 1e-05,
"loss": 2.2695,
"step": 874
},
{
"epoch": 1.4727540500736378,
"grad_norm": 0.12026406079530716,
"learning_rate": 1e-05,
"loss": 2.3313,
"step": 875
},
{
"epoch": 1.4744371975594361,
"grad_norm": 0.13317981362342834,
"learning_rate": 1e-05,
"loss": 2.209,
"step": 876
},
{
"epoch": 1.4761203450452345,
"grad_norm": 0.12904947996139526,
"learning_rate": 1e-05,
"loss": 2.2344,
"step": 877
},
{
"epoch": 1.477803492531033,
"grad_norm": 0.13126946985721588,
"learning_rate": 1e-05,
"loss": 2.2888,
"step": 878
},
{
"epoch": 1.4794866400168316,
"grad_norm": 0.128869891166687,
"learning_rate": 1e-05,
"loss": 2.1996,
"step": 879
},
{
"epoch": 1.48116978750263,
"grad_norm": 0.1279861181974411,
"learning_rate": 1e-05,
"loss": 2.1873,
"step": 880
},
{
"epoch": 1.4828529349884283,
"grad_norm": 0.11732237040996552,
"learning_rate": 1e-05,
"loss": 2.3259,
"step": 881
},
{
"epoch": 1.4845360824742269,
"grad_norm": 0.1279248595237732,
"learning_rate": 1e-05,
"loss": 2.386,
"step": 882
},
{
"epoch": 1.4862192299600252,
"grad_norm": 0.13578535616397858,
"learning_rate": 1e-05,
"loss": 2.2937,
"step": 883
},
{
"epoch": 1.4879023774458238,
"grad_norm": 0.13534606993198395,
"learning_rate": 1e-05,
"loss": 2.239,
"step": 884
},
{
"epoch": 1.489585524931622,
"grad_norm": 0.12359879165887833,
"learning_rate": 1e-05,
"loss": 2.3572,
"step": 885
},
{
"epoch": 1.4912686724174207,
"grad_norm": 0.1236250028014183,
"learning_rate": 1e-05,
"loss": 2.188,
"step": 886
},
{
"epoch": 1.492951819903219,
"grad_norm": 0.12695659697055817,
"learning_rate": 1e-05,
"loss": 2.2637,
"step": 887
},
{
"epoch": 1.4946349673890174,
"grad_norm": 0.1281343400478363,
"learning_rate": 1e-05,
"loss": 2.2961,
"step": 888
},
{
"epoch": 1.496318114874816,
"grad_norm": 0.12446150928735733,
"learning_rate": 1e-05,
"loss": 2.3362,
"step": 889
},
{
"epoch": 1.4980012623606145,
"grad_norm": 0.12564988434314728,
"learning_rate": 1e-05,
"loss": 2.288,
"step": 890
},
{
"epoch": 1.4996844098464128,
"grad_norm": 0.14049400389194489,
"learning_rate": 1e-05,
"loss": 2.2867,
"step": 891
},
{
"epoch": 1.5013675573322112,
"grad_norm": 0.12252961844205856,
"learning_rate": 1e-05,
"loss": 2.3511,
"step": 892
},
{
"epoch": 1.5030507048180097,
"grad_norm": 0.15993735194206238,
"learning_rate": 1e-05,
"loss": 2.0931,
"step": 893
},
{
"epoch": 1.504733852303808,
"grad_norm": 0.13673749566078186,
"learning_rate": 1e-05,
"loss": 2.2998,
"step": 894
},
{
"epoch": 1.5064169997896064,
"grad_norm": 0.11770147830247879,
"learning_rate": 1e-05,
"loss": 2.2883,
"step": 895
},
{
"epoch": 1.508100147275405,
"grad_norm": 0.11792504787445068,
"learning_rate": 1e-05,
"loss": 2.1893,
"step": 896
},
{
"epoch": 1.5097832947612035,
"grad_norm": 0.1405222862958908,
"learning_rate": 1e-05,
"loss": 2.2645,
"step": 897
},
{
"epoch": 1.5114664422470019,
"grad_norm": 0.1401311457157135,
"learning_rate": 1e-05,
"loss": 2.2085,
"step": 898
},
{
"epoch": 1.5131495897328002,
"grad_norm": 0.14068666100502014,
"learning_rate": 1e-05,
"loss": 2.2711,
"step": 899
},
{
"epoch": 1.5148327372185988,
"grad_norm": 0.12995976209640503,
"learning_rate": 1e-05,
"loss": 2.2883,
"step": 900
},
{
"epoch": 1.5165158847043974,
"grad_norm": 0.12454178184270859,
"learning_rate": 1e-05,
"loss": 2.2515,
"step": 901
},
{
"epoch": 1.5181990321901957,
"grad_norm": 0.12165191769599915,
"learning_rate": 1e-05,
"loss": 2.3621,
"step": 902
},
{
"epoch": 1.519882179675994,
"grad_norm": 0.1413601189851761,
"learning_rate": 1e-05,
"loss": 2.27,
"step": 903
},
{
"epoch": 1.5215653271617926,
"grad_norm": 0.13545894622802734,
"learning_rate": 1e-05,
"loss": 2.3008,
"step": 904
},
{
"epoch": 1.523248474647591,
"grad_norm": 0.12211872637271881,
"learning_rate": 1e-05,
"loss": 2.3921,
"step": 905
},
{
"epoch": 1.5249316221333893,
"grad_norm": 0.13053253293037415,
"learning_rate": 1e-05,
"loss": 2.2434,
"step": 906
},
{
"epoch": 1.5266147696191879,
"grad_norm": 0.12977124750614166,
"learning_rate": 1e-05,
"loss": 2.2366,
"step": 907
},
{
"epoch": 1.5282979171049864,
"grad_norm": 0.13451719284057617,
"learning_rate": 1e-05,
"loss": 2.3154,
"step": 908
},
{
"epoch": 1.5299810645907848,
"grad_norm": 0.11067184805870056,
"learning_rate": 1e-05,
"loss": 2.3296,
"step": 909
},
{
"epoch": 1.531664212076583,
"grad_norm": 0.12281223386526108,
"learning_rate": 1e-05,
"loss": 2.2479,
"step": 910
},
{
"epoch": 1.5333473595623817,
"grad_norm": 0.12240397185087204,
"learning_rate": 1e-05,
"loss": 2.3416,
"step": 911
},
{
"epoch": 1.5350305070481802,
"grad_norm": 0.14465166628360748,
"learning_rate": 1e-05,
"loss": 2.1801,
"step": 912
},
{
"epoch": 1.5367136545339786,
"grad_norm": 0.1263197958469391,
"learning_rate": 1e-05,
"loss": 2.2583,
"step": 913
},
{
"epoch": 1.538396802019777,
"grad_norm": 0.14653970301151276,
"learning_rate": 1e-05,
"loss": 2.2939,
"step": 914
},
{
"epoch": 1.5400799495055755,
"grad_norm": 0.1311267763376236,
"learning_rate": 1e-05,
"loss": 2.2517,
"step": 915
},
{
"epoch": 1.5417630969913738,
"grad_norm": 0.13173674046993256,
"learning_rate": 1e-05,
"loss": 2.309,
"step": 916
},
{
"epoch": 1.5434462444771722,
"grad_norm": 0.13140322268009186,
"learning_rate": 1e-05,
"loss": 2.1447,
"step": 917
},
{
"epoch": 1.5451293919629707,
"grad_norm": 0.12431302666664124,
"learning_rate": 1e-05,
"loss": 2.3315,
"step": 918
},
{
"epoch": 1.5468125394487693,
"grad_norm": 0.14358630776405334,
"learning_rate": 1e-05,
"loss": 2.2634,
"step": 919
},
{
"epoch": 1.5484956869345676,
"grad_norm": 0.1297353357076645,
"learning_rate": 1e-05,
"loss": 2.2489,
"step": 920
},
{
"epoch": 1.550178834420366,
"grad_norm": 0.12963449954986572,
"learning_rate": 1e-05,
"loss": 2.1533,
"step": 921
},
{
"epoch": 1.5518619819061645,
"grad_norm": 0.11558603495359421,
"learning_rate": 1e-05,
"loss": 2.2688,
"step": 922
},
{
"epoch": 1.553545129391963,
"grad_norm": 0.14222054183483124,
"learning_rate": 1e-05,
"loss": 2.2385,
"step": 923
},
{
"epoch": 1.5552282768777614,
"grad_norm": 0.1376868486404419,
"learning_rate": 1e-05,
"loss": 2.2051,
"step": 924
},
{
"epoch": 1.5569114243635598,
"grad_norm": 0.12993879616260529,
"learning_rate": 1e-05,
"loss": 2.3445,
"step": 925
},
{
"epoch": 1.5585945718493583,
"grad_norm": 0.14503213763237,
"learning_rate": 1e-05,
"loss": 2.215,
"step": 926
},
{
"epoch": 1.560277719335157,
"grad_norm": 0.1302722692489624,
"learning_rate": 1e-05,
"loss": 2.1945,
"step": 927
},
{
"epoch": 1.561960866820955,
"grad_norm": 0.13545845448970795,
"learning_rate": 1e-05,
"loss": 2.3059,
"step": 928
},
{
"epoch": 1.5636440143067536,
"grad_norm": 0.12279404699802399,
"learning_rate": 1e-05,
"loss": 2.3511,
"step": 929
},
{
"epoch": 1.5653271617925522,
"grad_norm": 0.13220550119876862,
"learning_rate": 1e-05,
"loss": 2.2837,
"step": 930
},
{
"epoch": 1.5670103092783505,
"grad_norm": 0.1407599151134491,
"learning_rate": 1e-05,
"loss": 2.2905,
"step": 931
},
{
"epoch": 1.5686934567641488,
"grad_norm": 0.12597431242465973,
"learning_rate": 1e-05,
"loss": 2.366,
"step": 932
},
{
"epoch": 1.5703766042499474,
"grad_norm": 0.12998835742473602,
"learning_rate": 1e-05,
"loss": 2.1067,
"step": 933
},
{
"epoch": 1.572059751735746,
"grad_norm": 0.14708921313285828,
"learning_rate": 1e-05,
"loss": 2.2687,
"step": 934
},
{
"epoch": 1.5737428992215443,
"grad_norm": 0.13333402574062347,
"learning_rate": 1e-05,
"loss": 2.3381,
"step": 935
},
{
"epoch": 1.5754260467073427,
"grad_norm": 0.14774633944034576,
"learning_rate": 1e-05,
"loss": 2.163,
"step": 936
},
{
"epoch": 1.5771091941931412,
"grad_norm": 0.1283462792634964,
"learning_rate": 1e-05,
"loss": 2.3892,
"step": 937
},
{
"epoch": 1.5787923416789398,
"grad_norm": 0.12011823058128357,
"learning_rate": 1e-05,
"loss": 2.2758,
"step": 938
},
{
"epoch": 1.580475489164738,
"grad_norm": 0.11618427187204361,
"learning_rate": 1e-05,
"loss": 2.2545,
"step": 939
},
{
"epoch": 1.5821586366505365,
"grad_norm": 0.12683863937854767,
"learning_rate": 1e-05,
"loss": 2.291,
"step": 940
},
{
"epoch": 1.583841784136335,
"grad_norm": 0.13158243894577026,
"learning_rate": 1e-05,
"loss": 2.3066,
"step": 941
},
{
"epoch": 1.5855249316221334,
"grad_norm": 0.13269281387329102,
"learning_rate": 1e-05,
"loss": 2.3442,
"step": 942
},
{
"epoch": 1.5872080791079317,
"grad_norm": 0.14047692716121674,
"learning_rate": 1e-05,
"loss": 2.3092,
"step": 943
},
{
"epoch": 1.5888912265937303,
"grad_norm": 0.1387140154838562,
"learning_rate": 1e-05,
"loss": 2.1482,
"step": 944
},
{
"epoch": 1.5905743740795288,
"grad_norm": 0.13907848298549652,
"learning_rate": 1e-05,
"loss": 2.3484,
"step": 945
},
{
"epoch": 1.5922575215653272,
"grad_norm": 0.13114407658576965,
"learning_rate": 1e-05,
"loss": 2.2195,
"step": 946
},
{
"epoch": 1.5939406690511255,
"grad_norm": 0.1368924379348755,
"learning_rate": 1e-05,
"loss": 2.322,
"step": 947
},
{
"epoch": 1.595623816536924,
"grad_norm": 0.141913041472435,
"learning_rate": 1e-05,
"loss": 2.2336,
"step": 948
},
{
"epoch": 1.5973069640227227,
"grad_norm": 0.13295848667621613,
"learning_rate": 1e-05,
"loss": 2.3081,
"step": 949
},
{
"epoch": 1.5989901115085208,
"grad_norm": 0.12306110560894012,
"learning_rate": 1e-05,
"loss": 2.3354,
"step": 950
},
{
"epoch": 1.6006732589943193,
"grad_norm": 0.12122649699449539,
"learning_rate": 1e-05,
"loss": 2.2839,
"step": 951
},
{
"epoch": 1.602356406480118,
"grad_norm": 0.13046576082706451,
"learning_rate": 1e-05,
"loss": 2.385,
"step": 952
},
{
"epoch": 1.6040395539659162,
"grad_norm": 0.1272476315498352,
"learning_rate": 1e-05,
"loss": 2.4153,
"step": 953
},
{
"epoch": 1.6057227014517146,
"grad_norm": 0.13073799014091492,
"learning_rate": 1e-05,
"loss": 2.2854,
"step": 954
},
{
"epoch": 1.6074058489375131,
"grad_norm": 0.12583526968955994,
"learning_rate": 1e-05,
"loss": 2.3318,
"step": 955
},
{
"epoch": 1.6090889964233117,
"grad_norm": 0.1474972665309906,
"learning_rate": 1e-05,
"loss": 2.2542,
"step": 956
},
{
"epoch": 1.61077214390911,
"grad_norm": 0.13445797562599182,
"learning_rate": 1e-05,
"loss": 2.3645,
"step": 957
},
{
"epoch": 1.6124552913949084,
"grad_norm": 0.13466110825538635,
"learning_rate": 1e-05,
"loss": 2.3394,
"step": 958
},
{
"epoch": 1.614138438880707,
"grad_norm": 0.13525816798210144,
"learning_rate": 1e-05,
"loss": 2.2471,
"step": 959
},
{
"epoch": 1.6158215863665055,
"grad_norm": 0.1377459019422531,
"learning_rate": 1e-05,
"loss": 2.2478,
"step": 960
},
{
"epoch": 1.6175047338523036,
"grad_norm": 0.1405583918094635,
"learning_rate": 1e-05,
"loss": 2.2146,
"step": 961
},
{
"epoch": 1.6191878813381022,
"grad_norm": 0.11743167042732239,
"learning_rate": 1e-05,
"loss": 2.3555,
"step": 962
},
{
"epoch": 1.6208710288239008,
"grad_norm": 0.13644517958164215,
"learning_rate": 1e-05,
"loss": 2.2155,
"step": 963
},
{
"epoch": 1.6225541763096991,
"grad_norm": 0.12609997391700745,
"learning_rate": 1e-05,
"loss": 2.2593,
"step": 964
},
{
"epoch": 1.6242373237954975,
"grad_norm": 0.13276560604572296,
"learning_rate": 1e-05,
"loss": 2.1737,
"step": 965
},
{
"epoch": 1.625920471281296,
"grad_norm": 0.13567714393138885,
"learning_rate": 1e-05,
"loss": 2.3336,
"step": 966
},
{
"epoch": 1.6276036187670946,
"grad_norm": 0.12559200823307037,
"learning_rate": 1e-05,
"loss": 2.3494,
"step": 967
},
{
"epoch": 1.629286766252893,
"grad_norm": 0.13090649247169495,
"learning_rate": 1e-05,
"loss": 2.1851,
"step": 968
},
{
"epoch": 1.6309699137386913,
"grad_norm": 0.15777987241744995,
"learning_rate": 1e-05,
"loss": 2.2205,
"step": 969
},
{
"epoch": 1.6326530612244898,
"grad_norm": 0.1433715522289276,
"learning_rate": 1e-05,
"loss": 2.2295,
"step": 970
},
{
"epoch": 1.6343362087102884,
"grad_norm": 0.1218508929014206,
"learning_rate": 1e-05,
"loss": 2.3762,
"step": 971
},
{
"epoch": 1.6360193561960865,
"grad_norm": 0.14540942013263702,
"learning_rate": 1e-05,
"loss": 2.2139,
"step": 972
},
{
"epoch": 1.637702503681885,
"grad_norm": 0.14829136431217194,
"learning_rate": 1e-05,
"loss": 2.2871,
"step": 973
},
{
"epoch": 1.6393856511676836,
"grad_norm": 0.12728969752788544,
"learning_rate": 1e-05,
"loss": 2.2917,
"step": 974
},
{
"epoch": 1.641068798653482,
"grad_norm": 0.1471221148967743,
"learning_rate": 1e-05,
"loss": 2.2012,
"step": 975
},
{
"epoch": 1.6427519461392803,
"grad_norm": 0.13320200145244598,
"learning_rate": 1e-05,
"loss": 2.2771,
"step": 976
},
{
"epoch": 1.644435093625079,
"grad_norm": 0.1363966464996338,
"learning_rate": 1e-05,
"loss": 2.3086,
"step": 977
},
{
"epoch": 1.6461182411108775,
"grad_norm": 0.13870568573474884,
"learning_rate": 1e-05,
"loss": 2.2898,
"step": 978
},
{
"epoch": 1.6478013885966758,
"grad_norm": 0.15152350068092346,
"learning_rate": 1e-05,
"loss": 2.2994,
"step": 979
},
{
"epoch": 1.6494845360824741,
"grad_norm": 0.13830937445163727,
"learning_rate": 1e-05,
"loss": 2.2108,
"step": 980
},
{
"epoch": 1.6511676835682727,
"grad_norm": 0.15544220805168152,
"learning_rate": 1e-05,
"loss": 2.4043,
"step": 981
},
{
"epoch": 1.6528508310540713,
"grad_norm": 0.13135483860969543,
"learning_rate": 1e-05,
"loss": 2.2373,
"step": 982
},
{
"epoch": 1.6545339785398696,
"grad_norm": 0.12355194985866547,
"learning_rate": 1e-05,
"loss": 2.4163,
"step": 983
},
{
"epoch": 1.656217126025668,
"grad_norm": 0.14110660552978516,
"learning_rate": 1e-05,
"loss": 2.2031,
"step": 984
},
{
"epoch": 1.6579002735114665,
"grad_norm": 0.13077346980571747,
"learning_rate": 1e-05,
"loss": 2.3601,
"step": 985
},
{
"epoch": 1.6595834209972649,
"grad_norm": 0.14212660491466522,
"learning_rate": 1e-05,
"loss": 2.197,
"step": 986
},
{
"epoch": 1.6612665684830632,
"grad_norm": 0.12336140871047974,
"learning_rate": 1e-05,
"loss": 2.4146,
"step": 987
},
{
"epoch": 1.6629497159688618,
"grad_norm": 0.15291054546833038,
"learning_rate": 1e-05,
"loss": 2.2764,
"step": 988
},
{
"epoch": 1.6646328634546603,
"grad_norm": 0.1272605061531067,
"learning_rate": 1e-05,
"loss": 2.2703,
"step": 989
},
{
"epoch": 1.6663160109404587,
"grad_norm": 0.13462689518928528,
"learning_rate": 1e-05,
"loss": 2.3188,
"step": 990
},
{
"epoch": 1.667999158426257,
"grad_norm": 0.13290910422801971,
"learning_rate": 1e-05,
"loss": 2.2172,
"step": 991
},
{
"epoch": 1.6696823059120556,
"grad_norm": 0.15105758607387543,
"learning_rate": 1e-05,
"loss": 2.2156,
"step": 992
},
{
"epoch": 1.6713654533978541,
"grad_norm": 0.13150456547737122,
"learning_rate": 1e-05,
"loss": 2.3362,
"step": 993
},
{
"epoch": 1.6730486008836525,
"grad_norm": 0.13139204680919647,
"learning_rate": 1e-05,
"loss": 2.3833,
"step": 994
},
{
"epoch": 1.6747317483694508,
"grad_norm": 0.14886420965194702,
"learning_rate": 1e-05,
"loss": 2.1893,
"step": 995
},
{
"epoch": 1.6764148958552494,
"grad_norm": 0.13227102160453796,
"learning_rate": 1e-05,
"loss": 2.4055,
"step": 996
},
{
"epoch": 1.6780980433410477,
"grad_norm": 0.12545333802700043,
"learning_rate": 1e-05,
"loss": 2.3311,
"step": 997
},
{
"epoch": 1.679781190826846,
"grad_norm": 0.13391169905662537,
"learning_rate": 1e-05,
"loss": 2.3022,
"step": 998
},
{
"epoch": 1.6814643383126446,
"grad_norm": 0.13013269007205963,
"learning_rate": 1e-05,
"loss": 2.2318,
"step": 999
},
{
"epoch": 1.6831474857984432,
"grad_norm": 0.1331031173467636,
"learning_rate": 1e-05,
"loss": 2.3022,
"step": 1000
},
{
"epoch": 1.6848306332842415,
"grad_norm": 0.14438873529434204,
"learning_rate": 1e-05,
"loss": 2.2388,
"step": 1001
},
{
"epoch": 1.6865137807700399,
"grad_norm": 0.1422380954027176,
"learning_rate": 1e-05,
"loss": 2.3145,
"step": 1002
},
{
"epoch": 1.6881969282558384,
"grad_norm": 0.13909044861793518,
"learning_rate": 1e-05,
"loss": 2.2249,
"step": 1003
},
{
"epoch": 1.689880075741637,
"grad_norm": 0.14147858321666718,
"learning_rate": 1e-05,
"loss": 2.3179,
"step": 1004
},
{
"epoch": 1.6915632232274354,
"grad_norm": 0.13203288614749908,
"learning_rate": 1e-05,
"loss": 2.1912,
"step": 1005
},
{
"epoch": 1.6932463707132337,
"grad_norm": 0.14461839199066162,
"learning_rate": 1e-05,
"loss": 2.1982,
"step": 1006
},
{
"epoch": 1.6949295181990323,
"grad_norm": 0.14539021253585815,
"learning_rate": 1e-05,
"loss": 2.2917,
"step": 1007
},
{
"epoch": 1.6966126656848306,
"grad_norm": 0.14774973690509796,
"learning_rate": 1e-05,
"loss": 2.2639,
"step": 1008
},
{
"epoch": 1.698295813170629,
"grad_norm": 0.14927157759666443,
"learning_rate": 1e-05,
"loss": 2.1956,
"step": 1009
},
{
"epoch": 1.6999789606564275,
"grad_norm": 0.1286613643169403,
"learning_rate": 1e-05,
"loss": 2.292,
"step": 1010
},
{
"epoch": 1.701662108142226,
"grad_norm": 0.12883049249649048,
"learning_rate": 1e-05,
"loss": 2.2573,
"step": 1011
},
{
"epoch": 1.7033452556280244,
"grad_norm": 0.14129754900932312,
"learning_rate": 1e-05,
"loss": 2.334,
"step": 1012
},
{
"epoch": 1.7050284031138228,
"grad_norm": 0.13216479122638702,
"learning_rate": 1e-05,
"loss": 2.2664,
"step": 1013
},
{
"epoch": 1.7067115505996213,
"grad_norm": 0.12611788511276245,
"learning_rate": 1e-05,
"loss": 2.3159,
"step": 1014
},
{
"epoch": 1.7083946980854199,
"grad_norm": 0.14012207090854645,
"learning_rate": 1e-05,
"loss": 2.4026,
"step": 1015
},
{
"epoch": 1.7100778455712182,
"grad_norm": 0.14449255168437958,
"learning_rate": 1e-05,
"loss": 2.3313,
"step": 1016
},
{
"epoch": 1.7117609930570166,
"grad_norm": 0.15093393623828888,
"learning_rate": 1e-05,
"loss": 2.2075,
"step": 1017
},
{
"epoch": 1.7134441405428151,
"grad_norm": 0.15169350802898407,
"learning_rate": 1e-05,
"loss": 2.1926,
"step": 1018
},
{
"epoch": 1.7151272880286135,
"grad_norm": 0.13613849878311157,
"learning_rate": 1e-05,
"loss": 2.3394,
"step": 1019
},
{
"epoch": 1.7168104355144118,
"grad_norm": 0.13525283336639404,
"learning_rate": 1e-05,
"loss": 2.2234,
"step": 1020
},
{
"epoch": 1.7184935830002104,
"grad_norm": 0.1529736965894699,
"learning_rate": 1e-05,
"loss": 2.1866,
"step": 1021
},
{
"epoch": 1.720176730486009,
"grad_norm": 0.13723863661289215,
"learning_rate": 1e-05,
"loss": 2.3027,
"step": 1022
},
{
"epoch": 1.7218598779718073,
"grad_norm": 0.16251115500926971,
"learning_rate": 1e-05,
"loss": 2.3428,
"step": 1023
},
{
"epoch": 1.7235430254576056,
"grad_norm": 0.1440790742635727,
"learning_rate": 1e-05,
"loss": 2.3298,
"step": 1024
},
{
"epoch": 1.7252261729434042,
"grad_norm": 0.13486018776893616,
"learning_rate": 1e-05,
"loss": 2.3826,
"step": 1025
},
{
"epoch": 1.7269093204292028,
"grad_norm": 0.15616028010845184,
"learning_rate": 1e-05,
"loss": 2.0817,
"step": 1026
},
{
"epoch": 1.728592467915001,
"grad_norm": 0.15306299924850464,
"learning_rate": 1e-05,
"loss": 2.2601,
"step": 1027
},
{
"epoch": 1.7302756154007994,
"grad_norm": 0.14421014487743378,
"learning_rate": 1e-05,
"loss": 2.1998,
"step": 1028
},
{
"epoch": 1.731958762886598,
"grad_norm": 0.14438478648662567,
"learning_rate": 1e-05,
"loss": 2.262,
"step": 1029
},
{
"epoch": 1.7336419103723963,
"grad_norm": 0.13325351476669312,
"learning_rate": 1e-05,
"loss": 2.2852,
"step": 1030
},
{
"epoch": 1.7353250578581947,
"grad_norm": 0.14232920110225677,
"learning_rate": 1e-05,
"loss": 2.3147,
"step": 1031
},
{
"epoch": 1.7370082053439933,
"grad_norm": 0.1394515186548233,
"learning_rate": 1e-05,
"loss": 2.2781,
"step": 1032
},
{
"epoch": 1.7386913528297918,
"grad_norm": 0.12838682532310486,
"learning_rate": 1e-05,
"loss": 2.2827,
"step": 1033
},
{
"epoch": 1.7403745003155902,
"grad_norm": 0.15612417459487915,
"learning_rate": 1e-05,
"loss": 2.3108,
"step": 1034
},
{
"epoch": 1.7420576478013885,
"grad_norm": 0.14740139245986938,
"learning_rate": 1e-05,
"loss": 2.2412,
"step": 1035
},
{
"epoch": 1.743740795287187,
"grad_norm": 0.1541980355978012,
"learning_rate": 1e-05,
"loss": 2.3156,
"step": 1036
},
{
"epoch": 1.7454239427729856,
"grad_norm": 0.14056488871574402,
"learning_rate": 1e-05,
"loss": 2.1829,
"step": 1037
},
{
"epoch": 1.747107090258784,
"grad_norm": 0.143393874168396,
"learning_rate": 1e-05,
"loss": 2.2717,
"step": 1038
},
{
"epoch": 1.7487902377445823,
"grad_norm": 0.14296631515026093,
"learning_rate": 1e-05,
"loss": 2.342,
"step": 1039
},
{
"epoch": 1.7504733852303809,
"grad_norm": 0.13753627240657806,
"learning_rate": 1e-05,
"loss": 2.324,
"step": 1040
},
{
"epoch": 1.7521565327161792,
"grad_norm": 0.13361461460590363,
"learning_rate": 1e-05,
"loss": 2.3549,
"step": 1041
},
{
"epoch": 1.7538396802019776,
"grad_norm": 0.16176526248455048,
"learning_rate": 1e-05,
"loss": 2.0996,
"step": 1042
},
{
"epoch": 1.7555228276877761,
"grad_norm": 0.14512574672698975,
"learning_rate": 1e-05,
"loss": 2.3289,
"step": 1043
},
{
"epoch": 1.7572059751735747,
"grad_norm": 0.14329467713832855,
"learning_rate": 1e-05,
"loss": 2.2429,
"step": 1044
},
{
"epoch": 1.758889122659373,
"grad_norm": 0.1415308713912964,
"learning_rate": 1e-05,
"loss": 2.2976,
"step": 1045
},
{
"epoch": 1.7605722701451714,
"grad_norm": 0.13017630577087402,
"learning_rate": 1e-05,
"loss": 2.3142,
"step": 1046
},
{
"epoch": 1.76225541763097,
"grad_norm": 0.14865103363990784,
"learning_rate": 1e-05,
"loss": 2.2659,
"step": 1047
},
{
"epoch": 1.7639385651167685,
"grad_norm": 0.13973674178123474,
"learning_rate": 1e-05,
"loss": 2.1975,
"step": 1048
},
{
"epoch": 1.7656217126025668,
"grad_norm": 0.12378077954053879,
"learning_rate": 1e-05,
"loss": 2.4469,
"step": 1049
},
{
"epoch": 1.7673048600883652,
"grad_norm": 0.13462629914283752,
"learning_rate": 1e-05,
"loss": 2.332,
"step": 1050
},
{
"epoch": 1.7689880075741637,
"grad_norm": 0.14375431835651398,
"learning_rate": 1e-05,
"loss": 2.2834,
"step": 1051
},
{
"epoch": 1.770671155059962,
"grad_norm": 0.1413864940404892,
"learning_rate": 1e-05,
"loss": 2.2769,
"step": 1052
},
{
"epoch": 1.7723543025457604,
"grad_norm": 0.15052342414855957,
"learning_rate": 1e-05,
"loss": 2.2522,
"step": 1053
},
{
"epoch": 1.774037450031559,
"grad_norm": 0.15616975724697113,
"learning_rate": 1e-05,
"loss": 2.1501,
"step": 1054
},
{
"epoch": 1.7757205975173576,
"grad_norm": 0.16257071495056152,
"learning_rate": 1e-05,
"loss": 2.1545,
"step": 1055
},
{
"epoch": 1.777403745003156,
"grad_norm": 0.13512100279331207,
"learning_rate": 1e-05,
"loss": 2.2218,
"step": 1056
},
{
"epoch": 1.7790868924889542,
"grad_norm": 0.1581428200006485,
"learning_rate": 1e-05,
"loss": 2.1865,
"step": 1057
},
{
"epoch": 1.7807700399747528,
"grad_norm": 0.13829343020915985,
"learning_rate": 1e-05,
"loss": 2.3337,
"step": 1058
},
{
"epoch": 1.7824531874605514,
"grad_norm": 0.16639141738414764,
"learning_rate": 1e-05,
"loss": 2.2325,
"step": 1059
},
{
"epoch": 1.7841363349463497,
"grad_norm": 0.1412006914615631,
"learning_rate": 1e-05,
"loss": 2.3389,
"step": 1060
},
{
"epoch": 1.785819482432148,
"grad_norm": 0.13130658864974976,
"learning_rate": 1e-05,
"loss": 2.3376,
"step": 1061
},
{
"epoch": 1.7875026299179466,
"grad_norm": 0.1495353728532791,
"learning_rate": 1e-05,
"loss": 2.2666,
"step": 1062
},
{
"epoch": 1.789185777403745,
"grad_norm": 0.15077506005764008,
"learning_rate": 1e-05,
"loss": 2.228,
"step": 1063
},
{
"epoch": 1.7908689248895433,
"grad_norm": 0.1426386535167694,
"learning_rate": 1e-05,
"loss": 2.2727,
"step": 1064
},
{
"epoch": 1.7925520723753419,
"grad_norm": 0.14268244802951813,
"learning_rate": 1e-05,
"loss": 2.3643,
"step": 1065
},
{
"epoch": 1.7942352198611404,
"grad_norm": 0.14923584461212158,
"learning_rate": 1e-05,
"loss": 2.333,
"step": 1066
},
{
"epoch": 1.7959183673469388,
"grad_norm": 0.15571311116218567,
"learning_rate": 1e-05,
"loss": 2.3171,
"step": 1067
},
{
"epoch": 1.7976015148327371,
"grad_norm": 0.13931907713413239,
"learning_rate": 1e-05,
"loss": 2.2164,
"step": 1068
},
{
"epoch": 1.7992846623185357,
"grad_norm": 0.1513443887233734,
"learning_rate": 1e-05,
"loss": 2.2885,
"step": 1069
},
{
"epoch": 1.8009678098043342,
"grad_norm": 0.14123128354549408,
"learning_rate": 1e-05,
"loss": 2.3517,
"step": 1070
},
{
"epoch": 1.8026509572901326,
"grad_norm": 0.16668306291103363,
"learning_rate": 1e-05,
"loss": 2.1907,
"step": 1071
},
{
"epoch": 1.804334104775931,
"grad_norm": 0.14049063622951508,
"learning_rate": 1e-05,
"loss": 2.4216,
"step": 1072
},
{
"epoch": 1.8060172522617295,
"grad_norm": 0.13806495070457458,
"learning_rate": 1e-05,
"loss": 2.3367,
"step": 1073
},
{
"epoch": 1.8077003997475278,
"grad_norm": 0.14562048017978668,
"learning_rate": 1e-05,
"loss": 2.2303,
"step": 1074
},
{
"epoch": 1.8093835472333262,
"grad_norm": 0.16803675889968872,
"learning_rate": 1e-05,
"loss": 2.2404,
"step": 1075
},
{
"epoch": 1.8110666947191247,
"grad_norm": 0.14971864223480225,
"learning_rate": 1e-05,
"loss": 2.1941,
"step": 1076
},
{
"epoch": 1.8127498422049233,
"grad_norm": 0.162116140127182,
"learning_rate": 1e-05,
"loss": 2.2034,
"step": 1077
},
{
"epoch": 1.8144329896907216,
"grad_norm": 0.1417408138513565,
"learning_rate": 1e-05,
"loss": 2.2991,
"step": 1078
},
{
"epoch": 1.81611613717652,
"grad_norm": 0.14334024488925934,
"learning_rate": 1e-05,
"loss": 2.3796,
"step": 1079
},
{
"epoch": 1.8177992846623185,
"grad_norm": 0.13600003719329834,
"learning_rate": 1e-05,
"loss": 2.2322,
"step": 1080
},
{
"epoch": 1.8194824321481171,
"grad_norm": 0.1557435244321823,
"learning_rate": 1e-05,
"loss": 2.2151,
"step": 1081
},
{
"epoch": 1.8211655796339155,
"grad_norm": 0.14444471895694733,
"learning_rate": 1e-05,
"loss": 2.2778,
"step": 1082
},
{
"epoch": 1.8228487271197138,
"grad_norm": 0.15237338840961456,
"learning_rate": 1e-05,
"loss": 2.1863,
"step": 1083
},
{
"epoch": 1.8245318746055124,
"grad_norm": 0.1488647758960724,
"learning_rate": 1e-05,
"loss": 2.1194,
"step": 1084
},
{
"epoch": 1.8262150220913107,
"grad_norm": 0.14532509446144104,
"learning_rate": 1e-05,
"loss": 2.3018,
"step": 1085
},
{
"epoch": 1.827898169577109,
"grad_norm": 0.1438300609588623,
"learning_rate": 1e-05,
"loss": 2.3542,
"step": 1086
},
{
"epoch": 1.8295813170629076,
"grad_norm": 0.13162897527217865,
"learning_rate": 1e-05,
"loss": 2.3762,
"step": 1087
},
{
"epoch": 1.8312644645487062,
"grad_norm": 0.14388734102249146,
"learning_rate": 1e-05,
"loss": 2.3097,
"step": 1088
},
{
"epoch": 1.8329476120345045,
"grad_norm": 0.1633898764848709,
"learning_rate": 1e-05,
"loss": 2.1975,
"step": 1089
},
{
"epoch": 1.8346307595203029,
"grad_norm": 0.14513400197029114,
"learning_rate": 1e-05,
"loss": 2.3562,
"step": 1090
},
{
"epoch": 1.8363139070061014,
"grad_norm": 0.1562061607837677,
"learning_rate": 1e-05,
"loss": 2.2384,
"step": 1091
},
{
"epoch": 1.8379970544919,
"grad_norm": 0.14833082258701324,
"learning_rate": 1e-05,
"loss": 2.199,
"step": 1092
},
{
"epoch": 1.8396802019776983,
"grad_norm": 0.14182843267917633,
"learning_rate": 1e-05,
"loss": 2.2632,
"step": 1093
},
{
"epoch": 1.8413633494634967,
"grad_norm": 0.16517210006713867,
"learning_rate": 1e-05,
"loss": 2.2719,
"step": 1094
},
{
"epoch": 1.8430464969492952,
"grad_norm": 0.1563366949558258,
"learning_rate": 1e-05,
"loss": 2.2285,
"step": 1095
},
{
"epoch": 1.8447296444350936,
"grad_norm": 0.1349581480026245,
"learning_rate": 1e-05,
"loss": 2.2998,
"step": 1096
},
{
"epoch": 1.846412791920892,
"grad_norm": 0.14647842943668365,
"learning_rate": 1e-05,
"loss": 2.2588,
"step": 1097
},
{
"epoch": 1.8480959394066905,
"grad_norm": 0.1527308076620102,
"learning_rate": 1e-05,
"loss": 2.1945,
"step": 1098
},
{
"epoch": 1.849779086892489,
"grad_norm": 0.16208425164222717,
"learning_rate": 1e-05,
"loss": 2.1692,
"step": 1099
},
{
"epoch": 1.8514622343782874,
"grad_norm": 0.15897248685359955,
"learning_rate": 1e-05,
"loss": 2.3582,
"step": 1100
},
{
"epoch": 1.8531453818640857,
"grad_norm": 0.14687612652778625,
"learning_rate": 1e-05,
"loss": 2.3057,
"step": 1101
},
{
"epoch": 1.8548285293498843,
"grad_norm": 0.1631488800048828,
"learning_rate": 1e-05,
"loss": 2.2521,
"step": 1102
},
{
"epoch": 1.8565116768356829,
"grad_norm": 0.14686156809329987,
"learning_rate": 1e-05,
"loss": 2.313,
"step": 1103
},
{
"epoch": 1.8581948243214812,
"grad_norm": 0.162966787815094,
"learning_rate": 1e-05,
"loss": 2.1968,
"step": 1104
},
{
"epoch": 1.8598779718072795,
"grad_norm": 0.15387648344039917,
"learning_rate": 1e-05,
"loss": 2.3059,
"step": 1105
},
{
"epoch": 1.861561119293078,
"grad_norm": 0.1489906907081604,
"learning_rate": 1e-05,
"loss": 2.2195,
"step": 1106
},
{
"epoch": 1.8632442667788764,
"grad_norm": 0.14351260662078857,
"learning_rate": 1e-05,
"loss": 2.2656,
"step": 1107
},
{
"epoch": 1.8649274142646748,
"grad_norm": 0.16010256111621857,
"learning_rate": 1e-05,
"loss": 2.3252,
"step": 1108
},
{
"epoch": 1.8666105617504734,
"grad_norm": 0.14475148916244507,
"learning_rate": 1e-05,
"loss": 2.2878,
"step": 1109
},
{
"epoch": 1.868293709236272,
"grad_norm": 0.14097367227077484,
"learning_rate": 1e-05,
"loss": 2.3716,
"step": 1110
},
{
"epoch": 1.8699768567220703,
"grad_norm": 0.15699978172779083,
"learning_rate": 1e-05,
"loss": 2.1678,
"step": 1111
},
{
"epoch": 1.8716600042078686,
"grad_norm": 0.1370065063238144,
"learning_rate": 1e-05,
"loss": 2.3315,
"step": 1112
},
{
"epoch": 1.8733431516936672,
"grad_norm": 0.1498231291770935,
"learning_rate": 1e-05,
"loss": 2.2949,
"step": 1113
},
{
"epoch": 1.8750262991794657,
"grad_norm": 0.13267523050308228,
"learning_rate": 1e-05,
"loss": 2.3535,
"step": 1114
},
{
"epoch": 1.876709446665264,
"grad_norm": 0.1453379988670349,
"learning_rate": 1e-05,
"loss": 2.2791,
"step": 1115
},
{
"epoch": 1.8783925941510624,
"grad_norm": 0.15499484539031982,
"learning_rate": 1e-05,
"loss": 2.2085,
"step": 1116
},
{
"epoch": 1.880075741636861,
"grad_norm": 0.14418251812458038,
"learning_rate": 1e-05,
"loss": 2.2793,
"step": 1117
},
{
"epoch": 1.8817588891226595,
"grad_norm": 0.13686548173427582,
"learning_rate": 1e-05,
"loss": 2.4175,
"step": 1118
},
{
"epoch": 1.8834420366084577,
"grad_norm": 0.17202888429164886,
"learning_rate": 1e-05,
"loss": 2.2196,
"step": 1119
},
{
"epoch": 1.8851251840942562,
"grad_norm": 0.1437048763036728,
"learning_rate": 1e-05,
"loss": 2.2688,
"step": 1120
},
{
"epoch": 1.8868083315800548,
"grad_norm": 0.13868288695812225,
"learning_rate": 1e-05,
"loss": 2.2971,
"step": 1121
},
{
"epoch": 1.8884914790658531,
"grad_norm": 0.133874773979187,
"learning_rate": 1e-05,
"loss": 2.3228,
"step": 1122
},
{
"epoch": 1.8901746265516515,
"grad_norm": 0.15967018902301788,
"learning_rate": 1e-05,
"loss": 2.2346,
"step": 1123
},
{
"epoch": 1.89185777403745,
"grad_norm": 0.15074019134044647,
"learning_rate": 1e-05,
"loss": 2.3577,
"step": 1124
},
{
"epoch": 1.8935409215232486,
"grad_norm": 0.13931475579738617,
"learning_rate": 1e-05,
"loss": 2.3789,
"step": 1125
},
{
"epoch": 1.895224069009047,
"grad_norm": 0.15354882180690765,
"learning_rate": 1e-05,
"loss": 2.184,
"step": 1126
},
{
"epoch": 1.8969072164948453,
"grad_norm": 0.15907764434814453,
"learning_rate": 1e-05,
"loss": 2.3638,
"step": 1127
},
{
"epoch": 1.8985903639806438,
"grad_norm": 0.13138049840927124,
"learning_rate": 1e-05,
"loss": 2.4543,
"step": 1128
},
{
"epoch": 1.9002735114664424,
"grad_norm": 0.14568856358528137,
"learning_rate": 1e-05,
"loss": 2.3064,
"step": 1129
},
{
"epoch": 1.9019566589522405,
"grad_norm": 0.1426182985305786,
"learning_rate": 1e-05,
"loss": 2.3223,
"step": 1130
},
{
"epoch": 1.903639806438039,
"grad_norm": 0.13313454389572144,
"learning_rate": 1e-05,
"loss": 2.3953,
"step": 1131
},
{
"epoch": 1.9053229539238377,
"grad_norm": 0.16987952589988708,
"learning_rate": 1e-05,
"loss": 2.1274,
"step": 1132
},
{
"epoch": 1.907006101409636,
"grad_norm": 0.1408863216638565,
"learning_rate": 1e-05,
"loss": 2.3242,
"step": 1133
},
{
"epoch": 1.9086892488954343,
"grad_norm": 0.14704225957393646,
"learning_rate": 1e-05,
"loss": 2.3687,
"step": 1134
},
{
"epoch": 1.910372396381233,
"grad_norm": 0.18410103023052216,
"learning_rate": 1e-05,
"loss": 2.1222,
"step": 1135
},
{
"epoch": 1.9120555438670315,
"grad_norm": 0.13889069855213165,
"learning_rate": 1e-05,
"loss": 2.3165,
"step": 1136
},
{
"epoch": 1.9137386913528298,
"grad_norm": 0.1532329022884369,
"learning_rate": 1e-05,
"loss": 2.2913,
"step": 1137
},
{
"epoch": 1.9154218388386282,
"grad_norm": 0.14806988835334778,
"learning_rate": 1e-05,
"loss": 2.2239,
"step": 1138
},
{
"epoch": 1.9171049863244267,
"grad_norm": 0.14964371919631958,
"learning_rate": 1e-05,
"loss": 2.2639,
"step": 1139
},
{
"epoch": 1.9187881338102253,
"grad_norm": 0.15137715637683868,
"learning_rate": 1e-05,
"loss": 2.3096,
"step": 1140
},
{
"epoch": 1.9204712812960234,
"grad_norm": 0.15892736613750458,
"learning_rate": 1e-05,
"loss": 2.3163,
"step": 1141
},
{
"epoch": 1.922154428781822,
"grad_norm": 0.15544387698173523,
"learning_rate": 1e-05,
"loss": 2.1825,
"step": 1142
},
{
"epoch": 1.9238375762676205,
"grad_norm": 0.14712852239608765,
"learning_rate": 1e-05,
"loss": 2.2659,
"step": 1143
},
{
"epoch": 1.9255207237534189,
"grad_norm": 0.1436305195093155,
"learning_rate": 1e-05,
"loss": 2.3101,
"step": 1144
},
{
"epoch": 1.9272038712392172,
"grad_norm": 0.16642406582832336,
"learning_rate": 1e-05,
"loss": 2.2156,
"step": 1145
},
{
"epoch": 1.9288870187250158,
"grad_norm": 0.16517338156700134,
"learning_rate": 1e-05,
"loss": 2.2561,
"step": 1146
},
{
"epoch": 1.9305701662108143,
"grad_norm": 0.1337500959634781,
"learning_rate": 1e-05,
"loss": 2.3818,
"step": 1147
},
{
"epoch": 1.9322533136966127,
"grad_norm": 0.15977586805820465,
"learning_rate": 1e-05,
"loss": 2.2377,
"step": 1148
},
{
"epoch": 1.933936461182411,
"grad_norm": 0.14951424300670624,
"learning_rate": 1e-05,
"loss": 2.2269,
"step": 1149
},
{
"epoch": 1.9356196086682096,
"grad_norm": 0.13450993597507477,
"learning_rate": 1e-05,
"loss": 2.3442,
"step": 1150
},
{
"epoch": 1.9373027561540082,
"grad_norm": 0.16469308733940125,
"learning_rate": 1e-05,
"loss": 2.3123,
"step": 1151
},
{
"epoch": 1.9389859036398063,
"grad_norm": 0.14135532081127167,
"learning_rate": 1e-05,
"loss": 2.387,
"step": 1152
},
{
"epoch": 1.9406690511256048,
"grad_norm": 0.13864876329898834,
"learning_rate": 1e-05,
"loss": 2.2661,
"step": 1153
},
{
"epoch": 1.9423521986114034,
"grad_norm": 0.16291983425617218,
"learning_rate": 1e-05,
"loss": 2.2617,
"step": 1154
},
{
"epoch": 1.9440353460972017,
"grad_norm": 0.13341820240020752,
"learning_rate": 1e-05,
"loss": 2.4299,
"step": 1155
},
{
"epoch": 1.945718493583,
"grad_norm": 0.15701517462730408,
"learning_rate": 1e-05,
"loss": 2.2211,
"step": 1156
},
{
"epoch": 1.9474016410687987,
"grad_norm": 0.16075365245342255,
"learning_rate": 1e-05,
"loss": 2.1801,
"step": 1157
},
{
"epoch": 1.9490847885545972,
"grad_norm": 0.15631234645843506,
"learning_rate": 1e-05,
"loss": 2.2152,
"step": 1158
},
{
"epoch": 1.9507679360403956,
"grad_norm": 0.16927126049995422,
"learning_rate": 1e-05,
"loss": 2.1776,
"step": 1159
},
{
"epoch": 1.952451083526194,
"grad_norm": 0.15192179381847382,
"learning_rate": 1e-05,
"loss": 2.2812,
"step": 1160
},
{
"epoch": 1.9541342310119925,
"grad_norm": 0.145833820104599,
"learning_rate": 1e-05,
"loss": 2.3124,
"step": 1161
},
{
"epoch": 1.955817378497791,
"grad_norm": 0.16952313482761383,
"learning_rate": 1e-05,
"loss": 2.1085,
"step": 1162
},
{
"epoch": 1.9575005259835891,
"grad_norm": 0.1629469394683838,
"learning_rate": 1e-05,
"loss": 2.2267,
"step": 1163
},
{
"epoch": 1.9591836734693877,
"grad_norm": 0.16672489047050476,
"learning_rate": 1e-05,
"loss": 2.3783,
"step": 1164
},
{
"epoch": 1.9608668209551863,
"grad_norm": 0.14810308814048767,
"learning_rate": 1e-05,
"loss": 2.3723,
"step": 1165
},
{
"epoch": 1.9625499684409846,
"grad_norm": 0.1435479074716568,
"learning_rate": 1e-05,
"loss": 2.2615,
"step": 1166
},
{
"epoch": 1.964233115926783,
"grad_norm": 0.149140864610672,
"learning_rate": 1e-05,
"loss": 2.2134,
"step": 1167
},
{
"epoch": 1.9659162634125815,
"grad_norm": 0.17785809934139252,
"learning_rate": 1e-05,
"loss": 2.1993,
"step": 1168
},
{
"epoch": 1.96759941089838,
"grad_norm": 0.15931861102581024,
"learning_rate": 1e-05,
"loss": 2.1807,
"step": 1169
},
{
"epoch": 1.9692825583841784,
"grad_norm": 0.16015268862247467,
"learning_rate": 1e-05,
"loss": 2.2737,
"step": 1170
},
{
"epoch": 1.9709657058699768,
"grad_norm": 0.14189362525939941,
"learning_rate": 1e-05,
"loss": 2.3416,
"step": 1171
},
{
"epoch": 1.9726488533557753,
"grad_norm": 0.1655077338218689,
"learning_rate": 1e-05,
"loss": 2.184,
"step": 1172
},
{
"epoch": 1.974332000841574,
"grad_norm": 0.17838408052921295,
"learning_rate": 1e-05,
"loss": 2.2466,
"step": 1173
},
{
"epoch": 1.9760151483273722,
"grad_norm": 0.16605247557163239,
"learning_rate": 1e-05,
"loss": 2.2019,
"step": 1174
},
{
"epoch": 1.9776982958131706,
"grad_norm": 0.15444627404212952,
"learning_rate": 1e-05,
"loss": 2.2382,
"step": 1175
},
{
"epoch": 1.9793814432989691,
"grad_norm": 0.15730591118335724,
"learning_rate": 1e-05,
"loss": 2.3335,
"step": 1176
},
{
"epoch": 1.9810645907847675,
"grad_norm": 0.17332051694393158,
"learning_rate": 1e-05,
"loss": 2.17,
"step": 1177
},
{
"epoch": 1.9827477382705658,
"grad_norm": 0.15129022300243378,
"learning_rate": 1e-05,
"loss": 2.2584,
"step": 1178
},
{
"epoch": 1.9844308857563644,
"grad_norm": 0.16302135586738586,
"learning_rate": 1e-05,
"loss": 2.1904,
"step": 1179
},
{
"epoch": 1.986114033242163,
"grad_norm": 0.14117322862148285,
"learning_rate": 1e-05,
"loss": 2.3611,
"step": 1180
},
{
"epoch": 1.9877971807279613,
"grad_norm": 0.14415599405765533,
"learning_rate": 1e-05,
"loss": 2.3503,
"step": 1181
},
{
"epoch": 1.9894803282137596,
"grad_norm": 0.15894141793251038,
"learning_rate": 1e-05,
"loss": 2.2253,
"step": 1182
},
{
"epoch": 1.9911634756995582,
"grad_norm": 0.15063215792179108,
"learning_rate": 1e-05,
"loss": 2.303,
"step": 1183
},
{
"epoch": 1.9928466231853568,
"grad_norm": 0.15843670070171356,
"learning_rate": 1e-05,
"loss": 2.2959,
"step": 1184
},
{
"epoch": 1.9945297706711551,
"grad_norm": 0.1457902193069458,
"learning_rate": 1e-05,
"loss": 2.3396,
"step": 1185
},
{
"epoch": 1.9962129181569535,
"grad_norm": 0.1694038361310959,
"learning_rate": 1e-05,
"loss": 2.3169,
"step": 1186
},
{
"epoch": 1.997896065642752,
"grad_norm": 0.16121593117713928,
"learning_rate": 1e-05,
"loss": 2.2754,
"step": 1187
},
{
"epoch": 1.9995792131285504,
"grad_norm": 0.16226674616336823,
"learning_rate": 1e-05,
"loss": 2.2498,
"step": 1188
},
{
"epoch": 1.9995792131285504,
"step": 1188,
"total_flos": 2.494777289795961e+18,
"train_loss": 2.3332799018834174,
"train_runtime": 81586.2049,
"train_samples_per_second": 0.932,
"train_steps_per_second": 0.015
}
],
"logging_steps": 1.0,
"max_steps": 1188,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.494777289795961e+18,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}