|
{
|
|
"best_metric": null,
|
|
"best_model_checkpoint": null,
|
|
"epoch": 34.89855072463768,
|
|
"eval_steps": 500,
|
|
"global_step": 1505,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.11594202898550725,
|
|
"grad_norm": 7.532149791717529,
|
|
"learning_rate": 4.999863832700438e-05,
|
|
"loss": 3.8116,
|
|
"num_input_tokens_seen": 106929,
|
|
"step": 5
|
|
},
|
|
{
|
|
"epoch": 0.2318840579710145,
|
|
"grad_norm": 4.141716480255127,
|
|
"learning_rate": 4.999455345634978e-05,
|
|
"loss": 3.6928,
|
|
"num_input_tokens_seen": 225964,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.34782608695652173,
|
|
"grad_norm": 3.470097780227661,
|
|
"learning_rate": 4.9987745833016855e-05,
|
|
"loss": 3.6227,
|
|
"num_input_tokens_seen": 362264,
|
|
"step": 15
|
|
},
|
|
{
|
|
"epoch": 0.463768115942029,
|
|
"grad_norm": 3.4544646739959717,
|
|
"learning_rate": 4.9978216198586135e-05,
|
|
"loss": 3.601,
|
|
"num_input_tokens_seen": 477807,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.5797101449275363,
|
|
"grad_norm": 3.249224901199341,
|
|
"learning_rate": 4.996596559115731e-05,
|
|
"loss": 3.539,
|
|
"num_input_tokens_seen": 588900,
|
|
"step": 25
|
|
},
|
|
{
|
|
"epoch": 0.6956521739130435,
|
|
"grad_norm": 3.395056962966919,
|
|
"learning_rate": 4.995099534523607e-05,
|
|
"loss": 3.4956,
|
|
"num_input_tokens_seen": 706077,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 0.8115942028985508,
|
|
"grad_norm": 3.997875213623047,
|
|
"learning_rate": 4.9933307091588796e-05,
|
|
"loss": 3.5044,
|
|
"num_input_tokens_seen": 853504,
|
|
"step": 35
|
|
},
|
|
{
|
|
"epoch": 0.927536231884058,
|
|
"grad_norm": 3.5168681144714355,
|
|
"learning_rate": 4.991290275706486e-05,
|
|
"loss": 3.4324,
|
|
"num_input_tokens_seen": 990472,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 1.0434782608695652,
|
|
"grad_norm": 7.144646167755127,
|
|
"learning_rate": 4.988978456438678e-05,
|
|
"loss": 3.2542,
|
|
"num_input_tokens_seen": 1125870,
|
|
"step": 45
|
|
},
|
|
{
|
|
"epoch": 1.1594202898550725,
|
|
"grad_norm": 3.257103681564331,
|
|
"learning_rate": 4.986395503190805e-05,
|
|
"loss": 2.9024,
|
|
"num_input_tokens_seen": 1249877,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 1.2753623188405796,
|
|
"grad_norm": 3.3208603858947754,
|
|
"learning_rate": 4.983541697333881e-05,
|
|
"loss": 2.8069,
|
|
"num_input_tokens_seen": 1375193,
|
|
"step": 55
|
|
},
|
|
{
|
|
"epoch": 1.391304347826087,
|
|
"grad_norm": 4.378167629241943,
|
|
"learning_rate": 4.980417349743936e-05,
|
|
"loss": 2.75,
|
|
"num_input_tokens_seen": 1489716,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 1.5072463768115942,
|
|
"grad_norm": 4.321849822998047,
|
|
"learning_rate": 4.9770228007681494e-05,
|
|
"loss": 2.7329,
|
|
"num_input_tokens_seen": 1600483,
|
|
"step": 65
|
|
},
|
|
{
|
|
"epoch": 1.6231884057971016,
|
|
"grad_norm": 3.6365067958831787,
|
|
"learning_rate": 4.973358420187776e-05,
|
|
"loss": 2.8212,
|
|
"num_input_tokens_seen": 1731315,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 1.7391304347826086,
|
|
"grad_norm": 3.74035906791687,
|
|
"learning_rate": 4.9694246071778604e-05,
|
|
"loss": 2.7935,
|
|
"num_input_tokens_seen": 1858269,
|
|
"step": 75
|
|
},
|
|
{
|
|
"epoch": 1.855072463768116,
|
|
"grad_norm": 23.26426124572754,
|
|
"learning_rate": 4.9652217902637596e-05,
|
|
"loss": 2.7305,
|
|
"num_input_tokens_seen": 1984587,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 1.971014492753623,
|
|
"grad_norm": 4.870578289031982,
|
|
"learning_rate": 4.9607504272744575e-05,
|
|
"loss": 2.6482,
|
|
"num_input_tokens_seen": 2109391,
|
|
"step": 85
|
|
},
|
|
{
|
|
"epoch": 2.0869565217391304,
|
|
"grad_norm": 4.419096946716309,
|
|
"learning_rate": 4.956011005292692e-05,
|
|
"loss": 2.4292,
|
|
"num_input_tokens_seen": 2246413,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 2.2028985507246377,
|
|
"grad_norm": 18.216915130615234,
|
|
"learning_rate": 4.951004040601898e-05,
|
|
"loss": 2.1416,
|
|
"num_input_tokens_seen": 2386890,
|
|
"step": 95
|
|
},
|
|
{
|
|
"epoch": 2.318840579710145,
|
|
"grad_norm": 4.4814581871032715,
|
|
"learning_rate": 4.945730078629964e-05,
|
|
"loss": 2.2847,
|
|
"num_input_tokens_seen": 2522302,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 2.4347826086956523,
|
|
"grad_norm": 406.9701232910156,
|
|
"learning_rate": 4.9401896938898185e-05,
|
|
"loss": 2.0944,
|
|
"num_input_tokens_seen": 2642208,
|
|
"step": 105
|
|
},
|
|
{
|
|
"epoch": 2.550724637681159,
|
|
"grad_norm": 3.5890581607818604,
|
|
"learning_rate": 4.934383489916843e-05,
|
|
"loss": 2.2862,
|
|
"num_input_tokens_seen": 2780587,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 2.6666666666666665,
|
|
"grad_norm": 5.334541320800781,
|
|
"learning_rate": 4.928312099203131e-05,
|
|
"loss": 2.105,
|
|
"num_input_tokens_seen": 2885320,
|
|
"step": 115
|
|
},
|
|
{
|
|
"epoch": 2.782608695652174,
|
|
"grad_norm": 5.464664936065674,
|
|
"learning_rate": 4.921976183128585e-05,
|
|
"loss": 2.0287,
|
|
"num_input_tokens_seen": 2996923,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 2.898550724637681,
|
|
"grad_norm": 4.113780975341797,
|
|
"learning_rate": 4.9153764318888706e-05,
|
|
"loss": 2.0162,
|
|
"num_input_tokens_seen": 3102391,
|
|
"step": 125
|
|
},
|
|
{
|
|
"epoch": 3.0144927536231885,
|
|
"grad_norm": 6.009971618652344,
|
|
"learning_rate": 4.908513564420231e-05,
|
|
"loss": 2.2464,
|
|
"num_input_tokens_seen": 3233443,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 3.130434782608696,
|
|
"grad_norm": 10.397327423095703,
|
|
"learning_rate": 4.90138832832117e-05,
|
|
"loss": 1.6561,
|
|
"num_input_tokens_seen": 3358966,
|
|
"step": 135
|
|
},
|
|
{
|
|
"epoch": 3.246376811594203,
|
|
"grad_norm": 4.9139556884765625,
|
|
"learning_rate": 4.894001499771015e-05,
|
|
"loss": 1.6113,
|
|
"num_input_tokens_seen": 3490069,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 3.36231884057971,
|
|
"grad_norm": 4.146034240722656,
|
|
"learning_rate": 4.886353883445363e-05,
|
|
"loss": 1.6235,
|
|
"num_input_tokens_seen": 3609842,
|
|
"step": 145
|
|
},
|
|
{
|
|
"epoch": 3.4782608695652173,
|
|
"grad_norm": 4.301880359649658,
|
|
"learning_rate": 4.878446312428424e-05,
|
|
"loss": 1.7873,
|
|
"num_input_tokens_seen": 3751570,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 3.5942028985507246,
|
|
"grad_norm": 3.9485158920288086,
|
|
"learning_rate": 4.8702796481222714e-05,
|
|
"loss": 1.3723,
|
|
"num_input_tokens_seen": 3865303,
|
|
"step": 155
|
|
},
|
|
{
|
|
"epoch": 3.710144927536232,
|
|
"grad_norm": 4.183668613433838,
|
|
"learning_rate": 4.861854780153004e-05,
|
|
"loss": 1.6512,
|
|
"num_input_tokens_seen": 3991347,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 3.8260869565217392,
|
|
"grad_norm": 5.1000471115112305,
|
|
"learning_rate": 4.853172626273841e-05,
|
|
"loss": 1.5524,
|
|
"num_input_tokens_seen": 4113654,
|
|
"step": 165
|
|
},
|
|
{
|
|
"epoch": 3.942028985507246,
|
|
"grad_norm": 4.142239570617676,
|
|
"learning_rate": 4.8442341322651385e-05,
|
|
"loss": 1.5954,
|
|
"num_input_tokens_seen": 4236348,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 4.057971014492754,
|
|
"grad_norm": 3.8976669311523438,
|
|
"learning_rate": 4.83504027183137e-05,
|
|
"loss": 1.1652,
|
|
"num_input_tokens_seen": 4340378,
|
|
"step": 175
|
|
},
|
|
{
|
|
"epoch": 4.173913043478261,
|
|
"grad_norm": 5.923389911651611,
|
|
"learning_rate": 4.825592046495054e-05,
|
|
"loss": 1.1995,
|
|
"num_input_tokens_seen": 4473601,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 4.2898550724637685,
|
|
"grad_norm": 4.220530033111572,
|
|
"learning_rate": 4.8158904854876555e-05,
|
|
"loss": 0.9431,
|
|
"num_input_tokens_seen": 4586911,
|
|
"step": 185
|
|
},
|
|
{
|
|
"epoch": 4.405797101449275,
|
|
"grad_norm": 5.896139144897461,
|
|
"learning_rate": 4.805936645637463e-05,
|
|
"loss": 1.1136,
|
|
"num_input_tokens_seen": 4702445,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 4.521739130434782,
|
|
"grad_norm": 4.467094421386719,
|
|
"learning_rate": 4.795731611254473e-05,
|
|
"loss": 1.1509,
|
|
"num_input_tokens_seen": 4831301,
|
|
"step": 195
|
|
},
|
|
{
|
|
"epoch": 4.63768115942029,
|
|
"grad_norm": 4.232386112213135,
|
|
"learning_rate": 4.785276494012263e-05,
|
|
"loss": 0.9962,
|
|
"num_input_tokens_seen": 4941656,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 4.753623188405797,
|
|
"grad_norm": 4.829892635345459,
|
|
"learning_rate": 4.7745724328269e-05,
|
|
"loss": 1.2377,
|
|
"num_input_tokens_seen": 5088437,
|
|
"step": 205
|
|
},
|
|
{
|
|
"epoch": 4.869565217391305,
|
|
"grad_norm": 4.1343913078308105,
|
|
"learning_rate": 4.763620593732867e-05,
|
|
"loss": 1.234,
|
|
"num_input_tokens_seen": 5219806,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 4.9855072463768115,
|
|
"grad_norm": 4.9217729568481445,
|
|
"learning_rate": 4.752422169756048e-05,
|
|
"loss": 1.1453,
|
|
"num_input_tokens_seen": 5340222,
|
|
"step": 215
|
|
},
|
|
{
|
|
"epoch": 5.101449275362318,
|
|
"grad_norm": 4.4605865478515625,
|
|
"learning_rate": 4.740978380783765e-05,
|
|
"loss": 0.9056,
|
|
"num_input_tokens_seen": 5476315,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 5.217391304347826,
|
|
"grad_norm": 4.396484375,
|
|
"learning_rate": 4.7292904734318924e-05,
|
|
"loss": 0.7349,
|
|
"num_input_tokens_seen": 5589951,
|
|
"step": 225
|
|
},
|
|
{
|
|
"epoch": 5.333333333333333,
|
|
"grad_norm": 4.053436279296875,
|
|
"learning_rate": 4.7173597209090534e-05,
|
|
"loss": 0.6968,
|
|
"num_input_tokens_seen": 5711449,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 5.449275362318841,
|
|
"grad_norm": 5.303736209869385,
|
|
"learning_rate": 4.70518742287793e-05,
|
|
"loss": 0.851,
|
|
"num_input_tokens_seen": 5852650,
|
|
"step": 235
|
|
},
|
|
{
|
|
"epoch": 5.565217391304348,
|
|
"grad_norm": 3.70810866355896,
|
|
"learning_rate": 4.6927749053136866e-05,
|
|
"loss": 0.716,
|
|
"num_input_tokens_seen": 5972289,
|
|
"step": 240
|
|
},
|
|
{
|
|
"epoch": 5.681159420289855,
|
|
"grad_norm": 3.9204599857330322,
|
|
"learning_rate": 4.6801235203595195e-05,
|
|
"loss": 0.6384,
|
|
"num_input_tokens_seen": 6088707,
|
|
"step": 245
|
|
},
|
|
{
|
|
"epoch": 5.797101449275362,
|
|
"grad_norm": 4.06931209564209,
|
|
"learning_rate": 4.667234646179368e-05,
|
|
"loss": 0.7799,
|
|
"num_input_tokens_seen": 6215471,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 5.913043478260869,
|
|
"grad_norm": 4.283618450164795,
|
|
"learning_rate": 4.654109686807787e-05,
|
|
"loss": 0.7923,
|
|
"num_input_tokens_seen": 6335935,
|
|
"step": 255
|
|
},
|
|
{
|
|
"epoch": 6.028985507246377,
|
|
"grad_norm": 4.719886302947998,
|
|
"learning_rate": 4.640750071996995e-05,
|
|
"loss": 0.7452,
|
|
"num_input_tokens_seen": 6463689,
|
|
"step": 260
|
|
},
|
|
{
|
|
"epoch": 6.144927536231884,
|
|
"grad_norm": 3.8415334224700928,
|
|
"learning_rate": 4.6271572570611296e-05,
|
|
"loss": 0.4085,
|
|
"num_input_tokens_seen": 6576954,
|
|
"step": 265
|
|
},
|
|
{
|
|
"epoch": 6.260869565217392,
|
|
"grad_norm": 4.19309663772583,
|
|
"learning_rate": 4.613332722717714e-05,
|
|
"loss": 0.5777,
|
|
"num_input_tokens_seen": 6714404,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 6.3768115942028984,
|
|
"grad_norm": 5.686235427856445,
|
|
"learning_rate": 4.5992779749263546e-05,
|
|
"loss": 0.4718,
|
|
"num_input_tokens_seen": 6840385,
|
|
"step": 275
|
|
},
|
|
{
|
|
"epoch": 6.492753623188406,
|
|
"grad_norm": 3.2365808486938477,
|
|
"learning_rate": 4.584994544724695e-05,
|
|
"loss": 0.3723,
|
|
"num_input_tokens_seen": 6954269,
|
|
"step": 280
|
|
},
|
|
{
|
|
"epoch": 6.608695652173913,
|
|
"grad_norm": 3.530801296234131,
|
|
"learning_rate": 4.5704839880616296e-05,
|
|
"loss": 0.4453,
|
|
"num_input_tokens_seen": 7076143,
|
|
"step": 285
|
|
},
|
|
{
|
|
"epoch": 6.72463768115942,
|
|
"grad_norm": 3.2134931087493896,
|
|
"learning_rate": 4.5557478856278114e-05,
|
|
"loss": 0.5742,
|
|
"num_input_tokens_seen": 7201833,
|
|
"step": 290
|
|
},
|
|
{
|
|
"epoch": 6.840579710144928,
|
|
"grad_norm": 6.281985282897949,
|
|
"learning_rate": 4.5407878426834596e-05,
|
|
"loss": 0.5291,
|
|
"num_input_tokens_seen": 7330479,
|
|
"step": 295
|
|
},
|
|
{
|
|
"epoch": 6.956521739130435,
|
|
"grad_norm": 17.072542190551758,
|
|
"learning_rate": 4.5256054888834934e-05,
|
|
"loss": 0.4968,
|
|
"num_input_tokens_seen": 7449244,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 7.072463768115942,
|
|
"grad_norm": 3.737456798553467,
|
|
"learning_rate": 4.5102024781000077e-05,
|
|
"loss": 0.421,
|
|
"num_input_tokens_seen": 7578947,
|
|
"step": 305
|
|
},
|
|
{
|
|
"epoch": 7.188405797101449,
|
|
"grad_norm": 2.6586523056030273,
|
|
"learning_rate": 4.4945804882421086e-05,
|
|
"loss": 0.2767,
|
|
"num_input_tokens_seen": 7691948,
|
|
"step": 310
|
|
},
|
|
{
|
|
"epoch": 7.304347826086957,
|
|
"grad_norm": 2.027702808380127,
|
|
"learning_rate": 4.478741221073136e-05,
|
|
"loss": 0.2922,
|
|
"num_input_tokens_seen": 7815786,
|
|
"step": 315
|
|
},
|
|
{
|
|
"epoch": 7.420289855072464,
|
|
"grad_norm": 3.4651787281036377,
|
|
"learning_rate": 4.4626864020252774e-05,
|
|
"loss": 0.2768,
|
|
"num_input_tokens_seen": 7925106,
|
|
"step": 320
|
|
},
|
|
{
|
|
"epoch": 7.536231884057971,
|
|
"grad_norm": 4.559577941894531,
|
|
"learning_rate": 4.446417780011618e-05,
|
|
"loss": 0.3281,
|
|
"num_input_tokens_seen": 8057202,
|
|
"step": 325
|
|
},
|
|
{
|
|
"epoch": 7.6521739130434785,
|
|
"grad_norm": 2.5885751247406006,
|
|
"learning_rate": 4.42993712723562e-05,
|
|
"loss": 0.3374,
|
|
"num_input_tokens_seen": 8187865,
|
|
"step": 330
|
|
},
|
|
{
|
|
"epoch": 7.768115942028985,
|
|
"grad_norm": 2.792222023010254,
|
|
"learning_rate": 4.413246238998069e-05,
|
|
"loss": 0.2491,
|
|
"num_input_tokens_seen": 8304605,
|
|
"step": 335
|
|
},
|
|
{
|
|
"epoch": 7.884057971014493,
|
|
"grad_norm": 3.610206127166748,
|
|
"learning_rate": 4.3963469335015085e-05,
|
|
"loss": 0.3893,
|
|
"num_input_tokens_seen": 8437319,
|
|
"step": 340
|
|
},
|
|
{
|
|
"epoch": 8.0,
|
|
"grad_norm": 4.31643533706665,
|
|
"learning_rate": 4.379241051652174e-05,
|
|
"loss": 0.3761,
|
|
"num_input_tokens_seen": 8573080,
|
|
"step": 345
|
|
},
|
|
{
|
|
"epoch": 8.115942028985508,
|
|
"grad_norm": 2.2834160327911377,
|
|
"learning_rate": 4.361930456859456e-05,
|
|
"loss": 0.236,
|
|
"num_input_tokens_seen": 8707741,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 8.231884057971014,
|
|
"grad_norm": 2.6929121017456055,
|
|
"learning_rate": 4.34441703483291e-05,
|
|
"loss": 0.1584,
|
|
"num_input_tokens_seen": 8825774,
|
|
"step": 355
|
|
},
|
|
{
|
|
"epoch": 8.347826086956522,
|
|
"grad_norm": 3.8095011711120605,
|
|
"learning_rate": 4.326702693376844e-05,
|
|
"loss": 0.1481,
|
|
"num_input_tokens_seen": 8932249,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 8.46376811594203,
|
|
"grad_norm": 2.6493489742279053,
|
|
"learning_rate": 4.308789362182492e-05,
|
|
"loss": 0.1743,
|
|
"num_input_tokens_seen": 9051548,
|
|
"step": 365
|
|
},
|
|
{
|
|
"epoch": 8.579710144927537,
|
|
"grad_norm": 30.796459197998047,
|
|
"learning_rate": 4.290678992617798e-05,
|
|
"loss": 0.3162,
|
|
"num_input_tokens_seen": 9197232,
|
|
"step": 370
|
|
},
|
|
{
|
|
"epoch": 8.695652173913043,
|
|
"grad_norm": 3.3164985179901123,
|
|
"learning_rate": 4.272373557514858e-05,
|
|
"loss": 0.2235,
|
|
"num_input_tokens_seen": 9317650,
|
|
"step": 375
|
|
},
|
|
{
|
|
"epoch": 8.81159420289855,
|
|
"grad_norm": 3.1515417098999023,
|
|
"learning_rate": 4.2538750509550054e-05,
|
|
"loss": 0.2504,
|
|
"num_input_tokens_seen": 9450765,
|
|
"step": 380
|
|
},
|
|
{
|
|
"epoch": 8.927536231884059,
|
|
"grad_norm": 3.3926901817321777,
|
|
"learning_rate": 4.235185488051585e-05,
|
|
"loss": 0.2136,
|
|
"num_input_tokens_seen": 9582961,
|
|
"step": 385
|
|
},
|
|
{
|
|
"epoch": 9.043478260869565,
|
|
"grad_norm": 4.670753002166748,
|
|
"learning_rate": 4.216306904730447e-05,
|
|
"loss": 0.1047,
|
|
"num_input_tokens_seen": 9678616,
|
|
"step": 390
|
|
},
|
|
{
|
|
"epoch": 9.159420289855072,
|
|
"grad_norm": 2.166652202606201,
|
|
"learning_rate": 4.1972413575081595e-05,
|
|
"loss": 0.1015,
|
|
"num_input_tokens_seen": 9788512,
|
|
"step": 395
|
|
},
|
|
{
|
|
"epoch": 9.27536231884058,
|
|
"grad_norm": 2.1161272525787354,
|
|
"learning_rate": 4.177990923267986e-05,
|
|
"loss": 0.1505,
|
|
"num_input_tokens_seen": 9916229,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 9.391304347826088,
|
|
"grad_norm": 2.378105401992798,
|
|
"learning_rate": 4.158557699033644e-05,
|
|
"loss": 0.1135,
|
|
"num_input_tokens_seen": 10042697,
|
|
"step": 405
|
|
},
|
|
{
|
|
"epoch": 9.507246376811594,
|
|
"grad_norm": 2.5567331314086914,
|
|
"learning_rate": 4.138943801740865e-05,
|
|
"loss": 0.1832,
|
|
"num_input_tokens_seen": 10171849,
|
|
"step": 410
|
|
},
|
|
{
|
|
"epoch": 9.623188405797102,
|
|
"grad_norm": 2.022610902786255,
|
|
"learning_rate": 4.119151368006793e-05,
|
|
"loss": 0.1178,
|
|
"num_input_tokens_seen": 10281924,
|
|
"step": 415
|
|
},
|
|
{
|
|
"epoch": 9.73913043478261,
|
|
"grad_norm": 2.5578079223632812,
|
|
"learning_rate": 4.099182553897229e-05,
|
|
"loss": 0.1426,
|
|
"num_input_tokens_seen": 10418758,
|
|
"step": 420
|
|
},
|
|
{
|
|
"epoch": 9.855072463768115,
|
|
"grad_norm": 2.7287228107452393,
|
|
"learning_rate": 4.079039534691767e-05,
|
|
"loss": 0.1603,
|
|
"num_input_tokens_seen": 10558322,
|
|
"step": 425
|
|
},
|
|
{
|
|
"epoch": 9.971014492753623,
|
|
"grad_norm": 2.361532688140869,
|
|
"learning_rate": 4.058724504646834e-05,
|
|
"loss": 0.1548,
|
|
"num_input_tokens_seen": 10679536,
|
|
"step": 430
|
|
},
|
|
{
|
|
"epoch": 10.08695652173913,
|
|
"grad_norm": 1.8757002353668213,
|
|
"learning_rate": 4.0382396767566536e-05,
|
|
"loss": 0.1407,
|
|
"num_input_tokens_seen": 10821076,
|
|
"step": 435
|
|
},
|
|
{
|
|
"epoch": 10.202898550724637,
|
|
"grad_norm": 2.352725028991699,
|
|
"learning_rate": 4.017587282512181e-05,
|
|
"loss": 0.0791,
|
|
"num_input_tokens_seen": 10949771,
|
|
"step": 440
|
|
},
|
|
{
|
|
"epoch": 10.318840579710145,
|
|
"grad_norm": 1.7948939800262451,
|
|
"learning_rate": 3.9967695716580224e-05,
|
|
"loss": 0.0722,
|
|
"num_input_tokens_seen": 11072044,
|
|
"step": 445
|
|
},
|
|
{
|
|
"epoch": 10.434782608695652,
|
|
"grad_norm": 1.954727292060852,
|
|
"learning_rate": 3.975788811947351e-05,
|
|
"loss": 0.0655,
|
|
"num_input_tokens_seen": 11182627,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 10.55072463768116,
|
|
"grad_norm": 2.143941640853882,
|
|
"learning_rate": 3.954647288894883e-05,
|
|
"loss": 0.0723,
|
|
"num_input_tokens_seen": 11303028,
|
|
"step": 455
|
|
},
|
|
{
|
|
"epoch": 10.666666666666666,
|
|
"grad_norm": 2.0527164936065674,
|
|
"learning_rate": 3.933347305527898e-05,
|
|
"loss": 0.0655,
|
|
"num_input_tokens_seen": 11415868,
|
|
"step": 460
|
|
},
|
|
{
|
|
"epoch": 10.782608695652174,
|
|
"grad_norm": 1.6390535831451416,
|
|
"learning_rate": 3.911891182135371e-05,
|
|
"loss": 0.1534,
|
|
"num_input_tokens_seen": 11555653,
|
|
"step": 465
|
|
},
|
|
{
|
|
"epoch": 10.898550724637682,
|
|
"grad_norm": 2.3848719596862793,
|
|
"learning_rate": 3.8902812560152066e-05,
|
|
"loss": 0.0947,
|
|
"num_input_tokens_seen": 11681065,
|
|
"step": 470
|
|
},
|
|
{
|
|
"epoch": 11.014492753623188,
|
|
"grad_norm": 2.2094757556915283,
|
|
"learning_rate": 3.868519881219631e-05,
|
|
"loss": 0.0868,
|
|
"num_input_tokens_seen": 11809957,
|
|
"step": 475
|
|
},
|
|
{
|
|
"epoch": 11.130434782608695,
|
|
"grad_norm": 4.137216567993164,
|
|
"learning_rate": 3.846609428298757e-05,
|
|
"loss": 0.0467,
|
|
"num_input_tokens_seen": 11937881,
|
|
"step": 480
|
|
},
|
|
{
|
|
"epoch": 11.246376811594203,
|
|
"grad_norm": 1.6658189296722412,
|
|
"learning_rate": 3.824552284042351e-05,
|
|
"loss": 0.0521,
|
|
"num_input_tokens_seen": 12048905,
|
|
"step": 485
|
|
},
|
|
{
|
|
"epoch": 11.36231884057971,
|
|
"grad_norm": 1.5732171535491943,
|
|
"learning_rate": 3.8023508512198256e-05,
|
|
"loss": 0.051,
|
|
"num_input_tokens_seen": 12185453,
|
|
"step": 490
|
|
},
|
|
{
|
|
"epoch": 11.478260869565217,
|
|
"grad_norm": 1.8459701538085938,
|
|
"learning_rate": 3.780007548318507e-05,
|
|
"loss": 0.0753,
|
|
"num_input_tokens_seen": 12310911,
|
|
"step": 495
|
|
},
|
|
{
|
|
"epoch": 11.594202898550725,
|
|
"grad_norm": 1.4724109172821045,
|
|
"learning_rate": 3.7575248092801686e-05,
|
|
"loss": 0.0601,
|
|
"num_input_tokens_seen": 12439708,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 11.710144927536232,
|
|
"grad_norm": 2.4690322875976562,
|
|
"learning_rate": 3.734905083235901e-05,
|
|
"loss": 0.0533,
|
|
"num_input_tokens_seen": 12554467,
|
|
"step": 505
|
|
},
|
|
{
|
|
"epoch": 11.826086956521738,
|
|
"grad_norm": 2.369218111038208,
|
|
"learning_rate": 3.712150834239313e-05,
|
|
"loss": 0.064,
|
|
"num_input_tokens_seen": 12682329,
|
|
"step": 510
|
|
},
|
|
{
|
|
"epoch": 11.942028985507246,
|
|
"grad_norm": 1.6901100873947144,
|
|
"learning_rate": 3.689264540998116e-05,
|
|
"loss": 0.0755,
|
|
"num_input_tokens_seen": 12800852,
|
|
"step": 515
|
|
},
|
|
{
|
|
"epoch": 12.057971014492754,
|
|
"grad_norm": 1.303114414215088,
|
|
"learning_rate": 3.66624869660411e-05,
|
|
"loss": 0.0553,
|
|
"num_input_tokens_seen": 12917527,
|
|
"step": 520
|
|
},
|
|
{
|
|
"epoch": 12.173913043478262,
|
|
"grad_norm": 1.1986353397369385,
|
|
"learning_rate": 3.6431058082615964e-05,
|
|
"loss": 0.0355,
|
|
"num_input_tokens_seen": 13044774,
|
|
"step": 525
|
|
},
|
|
{
|
|
"epoch": 12.289855072463768,
|
|
"grad_norm": 1.5653026103973389,
|
|
"learning_rate": 3.619838397014263e-05,
|
|
"loss": 0.0413,
|
|
"num_input_tokens_seen": 13175692,
|
|
"step": 530
|
|
},
|
|
{
|
|
"epoch": 12.405797101449275,
|
|
"grad_norm": 1.0767664909362793,
|
|
"learning_rate": 3.5964489974705553e-05,
|
|
"loss": 0.0596,
|
|
"num_input_tokens_seen": 13293164,
|
|
"step": 535
|
|
},
|
|
{
|
|
"epoch": 12.521739130434783,
|
|
"grad_norm": 1.6005312204360962,
|
|
"learning_rate": 3.572940157527572e-05,
|
|
"loss": 0.0479,
|
|
"num_input_tokens_seen": 13417894,
|
|
"step": 540
|
|
},
|
|
{
|
|
"epoch": 12.63768115942029,
|
|
"grad_norm": 1.627121925354004,
|
|
"learning_rate": 3.549314438093515e-05,
|
|
"loss": 0.047,
|
|
"num_input_tokens_seen": 13551913,
|
|
"step": 545
|
|
},
|
|
{
|
|
"epoch": 12.753623188405797,
|
|
"grad_norm": 2.239276647567749,
|
|
"learning_rate": 3.525574412808717e-05,
|
|
"loss": 0.0492,
|
|
"num_input_tokens_seen": 13675309,
|
|
"step": 550
|
|
},
|
|
{
|
|
"epoch": 12.869565217391305,
|
|
"grad_norm": 1.5702998638153076,
|
|
"learning_rate": 3.501722667765286e-05,
|
|
"loss": 0.0471,
|
|
"num_input_tokens_seen": 13797691,
|
|
"step": 555
|
|
},
|
|
{
|
|
"epoch": 12.985507246376812,
|
|
"grad_norm": 1.8216972351074219,
|
|
"learning_rate": 3.47776180122539e-05,
|
|
"loss": 0.1041,
|
|
"num_input_tokens_seen": 13919770,
|
|
"step": 560
|
|
},
|
|
{
|
|
"epoch": 13.101449275362318,
|
|
"grad_norm": 0.9026144742965698,
|
|
"learning_rate": 3.453694423338225e-05,
|
|
"loss": 0.0282,
|
|
"num_input_tokens_seen": 14037673,
|
|
"step": 565
|
|
},
|
|
{
|
|
"epoch": 13.217391304347826,
|
|
"grad_norm": 1.4504765272140503,
|
|
"learning_rate": 3.4295231558556715e-05,
|
|
"loss": 0.0272,
|
|
"num_input_tokens_seen": 14167090,
|
|
"step": 570
|
|
},
|
|
{
|
|
"epoch": 13.333333333333334,
|
|
"grad_norm": 1.4278969764709473,
|
|
"learning_rate": 3.4052506318467084e-05,
|
|
"loss": 0.0342,
|
|
"num_input_tokens_seen": 14311710,
|
|
"step": 575
|
|
},
|
|
{
|
|
"epoch": 13.44927536231884,
|
|
"grad_norm": 1.1284997463226318,
|
|
"learning_rate": 3.3808794954105716e-05,
|
|
"loss": 0.0855,
|
|
"num_input_tokens_seen": 14404322,
|
|
"step": 580
|
|
},
|
|
{
|
|
"epoch": 13.565217391304348,
|
|
"grad_norm": 1.4915614128112793,
|
|
"learning_rate": 3.356412401388732e-05,
|
|
"loss": 0.0378,
|
|
"num_input_tokens_seen": 14530794,
|
|
"step": 585
|
|
},
|
|
{
|
|
"epoch": 13.681159420289855,
|
|
"grad_norm": 1.372157096862793,
|
|
"learning_rate": 3.3318520150756846e-05,
|
|
"loss": 0.0457,
|
|
"num_input_tokens_seen": 14637342,
|
|
"step": 590
|
|
},
|
|
{
|
|
"epoch": 13.797101449275363,
|
|
"grad_norm": 1.6492116451263428,
|
|
"learning_rate": 3.307201011928616e-05,
|
|
"loss": 0.0453,
|
|
"num_input_tokens_seen": 14787534,
|
|
"step": 595
|
|
},
|
|
{
|
|
"epoch": 13.91304347826087,
|
|
"grad_norm": 1.3583859205245972,
|
|
"learning_rate": 3.282462077275947e-05,
|
|
"loss": 0.0378,
|
|
"num_input_tokens_seen": 14909175,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 14.028985507246377,
|
|
"grad_norm": 1.0751795768737793,
|
|
"learning_rate": 3.257637906024822e-05,
|
|
"loss": 0.0296,
|
|
"num_input_tokens_seen": 15030530,
|
|
"step": 605
|
|
},
|
|
{
|
|
"epoch": 14.144927536231885,
|
|
"grad_norm": 1.474602222442627,
|
|
"learning_rate": 3.2327312023675287e-05,
|
|
"loss": 0.0216,
|
|
"num_input_tokens_seen": 15148359,
|
|
"step": 610
|
|
},
|
|
{
|
|
"epoch": 14.26086956521739,
|
|
"grad_norm": 1.0749961137771606,
|
|
"learning_rate": 3.2077446794869295e-05,
|
|
"loss": 0.0299,
|
|
"num_input_tokens_seen": 15280749,
|
|
"step": 615
|
|
},
|
|
{
|
|
"epoch": 14.376811594202898,
|
|
"grad_norm": 1.4042794704437256,
|
|
"learning_rate": 3.1826810592609036e-05,
|
|
"loss": 0.0247,
|
|
"num_input_tokens_seen": 15397167,
|
|
"step": 620
|
|
},
|
|
{
|
|
"epoch": 14.492753623188406,
|
|
"grad_norm": 1.2280118465423584,
|
|
"learning_rate": 3.157543071965835e-05,
|
|
"loss": 0.0455,
|
|
"num_input_tokens_seen": 15522794,
|
|
"step": 625
|
|
},
|
|
{
|
|
"epoch": 14.608695652173914,
|
|
"grad_norm": 1.2819784879684448,
|
|
"learning_rate": 3.132333455979202e-05,
|
|
"loss": 0.0262,
|
|
"num_input_tokens_seen": 15637987,
|
|
"step": 630
|
|
},
|
|
{
|
|
"epoch": 14.72463768115942,
|
|
"grad_norm": 1.2691748142242432,
|
|
"learning_rate": 3.107054957481271e-05,
|
|
"loss": 0.0281,
|
|
"num_input_tokens_seen": 15773163,
|
|
"step": 635
|
|
},
|
|
{
|
|
"epoch": 14.840579710144928,
|
|
"grad_norm": 1.2752504348754883,
|
|
"learning_rate": 3.081710330155942e-05,
|
|
"loss": 0.0294,
|
|
"num_input_tokens_seen": 15892659,
|
|
"step": 640
|
|
},
|
|
{
|
|
"epoch": 14.956521739130435,
|
|
"grad_norm": 1.3479197025299072,
|
|
"learning_rate": 3.056302334890786e-05,
|
|
"loss": 0.0291,
|
|
"num_input_tokens_seen": 16024576,
|
|
"step": 645
|
|
},
|
|
{
|
|
"epoch": 15.072463768115941,
|
|
"grad_norm": 1.3151382207870483,
|
|
"learning_rate": 3.030833739476285e-05,
|
|
"loss": 0.0216,
|
|
"num_input_tokens_seen": 16151987,
|
|
"step": 650
|
|
},
|
|
{
|
|
"epoch": 15.18840579710145,
|
|
"grad_norm": 2.3882877826690674,
|
|
"learning_rate": 3.0053073183043256e-05,
|
|
"loss": 0.0218,
|
|
"num_input_tokens_seen": 16278639,
|
|
"step": 655
|
|
},
|
|
{
|
|
"epoch": 15.304347826086957,
|
|
"grad_norm": 0.9794278144836426,
|
|
"learning_rate": 2.979725852065981e-05,
|
|
"loss": 0.0283,
|
|
"num_input_tokens_seen": 16414743,
|
|
"step": 660
|
|
},
|
|
{
|
|
"epoch": 15.420289855072463,
|
|
"grad_norm": 0.8964869976043701,
|
|
"learning_rate": 2.954092127448591e-05,
|
|
"loss": 0.0259,
|
|
"num_input_tokens_seen": 16529298,
|
|
"step": 665
|
|
},
|
|
{
|
|
"epoch": 15.53623188405797,
|
|
"grad_norm": 1.1441810131072998,
|
|
"learning_rate": 2.9284089368322045e-05,
|
|
"loss": 0.0716,
|
|
"num_input_tokens_seen": 16655909,
|
|
"step": 670
|
|
},
|
|
{
|
|
"epoch": 15.652173913043478,
|
|
"grad_norm": 1.0959213972091675,
|
|
"learning_rate": 2.9026790779853874e-05,
|
|
"loss": 0.025,
|
|
"num_input_tokens_seen": 16798263,
|
|
"step": 675
|
|
},
|
|
{
|
|
"epoch": 15.768115942028986,
|
|
"grad_norm": 1.0119343996047974,
|
|
"learning_rate": 2.876905353760459e-05,
|
|
"loss": 0.0218,
|
|
"num_input_tokens_seen": 16916827,
|
|
"step": 680
|
|
},
|
|
{
|
|
"epoch": 15.884057971014492,
|
|
"grad_norm": 1.1373978853225708,
|
|
"learning_rate": 2.8510905717881614e-05,
|
|
"loss": 0.0231,
|
|
"num_input_tokens_seen": 17040247,
|
|
"step": 685
|
|
},
|
|
{
|
|
"epoch": 16.0,
|
|
"grad_norm": 1.2512497901916504,
|
|
"learning_rate": 2.8252375441718137e-05,
|
|
"loss": 0.0228,
|
|
"num_input_tokens_seen": 17146160,
|
|
"step": 690
|
|
},
|
|
{
|
|
"epoch": 16.115942028985508,
|
|
"grad_norm": 0.7410117387771606,
|
|
"learning_rate": 2.7993490871809808e-05,
|
|
"loss": 0.029,
|
|
"num_input_tokens_seen": 17284643,
|
|
"step": 695
|
|
},
|
|
{
|
|
"epoch": 16.231884057971016,
|
|
"grad_norm": 1.0934263467788696,
|
|
"learning_rate": 2.7734280209446865e-05,
|
|
"loss": 0.0199,
|
|
"num_input_tokens_seen": 17426644,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 16.347826086956523,
|
|
"grad_norm": 1.0034395456314087,
|
|
"learning_rate": 2.7474771691442018e-05,
|
|
"loss": 0.0259,
|
|
"num_input_tokens_seen": 17541812,
|
|
"step": 705
|
|
},
|
|
{
|
|
"epoch": 16.463768115942027,
|
|
"grad_norm": 1.4287781715393066,
|
|
"learning_rate": 2.721499358705458e-05,
|
|
"loss": 0.021,
|
|
"num_input_tokens_seen": 17667755,
|
|
"step": 710
|
|
},
|
|
{
|
|
"epoch": 16.579710144927535,
|
|
"grad_norm": 1.0989606380462646,
|
|
"learning_rate": 2.6954974194910888e-05,
|
|
"loss": 0.0199,
|
|
"num_input_tokens_seen": 17788162,
|
|
"step": 715
|
|
},
|
|
{
|
|
"epoch": 16.695652173913043,
|
|
"grad_norm": 0.9687130451202393,
|
|
"learning_rate": 2.6694741839921732e-05,
|
|
"loss": 0.0189,
|
|
"num_input_tokens_seen": 17911718,
|
|
"step": 720
|
|
},
|
|
{
|
|
"epoch": 16.81159420289855,
|
|
"grad_norm": 1.143617033958435,
|
|
"learning_rate": 2.6434324870196748e-05,
|
|
"loss": 0.0169,
|
|
"num_input_tokens_seen": 18018729,
|
|
"step": 725
|
|
},
|
|
{
|
|
"epoch": 16.92753623188406,
|
|
"grad_norm": 1.1395140886306763,
|
|
"learning_rate": 2.617375165395634e-05,
|
|
"loss": 0.0209,
|
|
"num_input_tokens_seen": 18139681,
|
|
"step": 730
|
|
},
|
|
{
|
|
"epoch": 17.043478260869566,
|
|
"grad_norm": 0.881986677646637,
|
|
"learning_rate": 2.5913050576441477e-05,
|
|
"loss": 0.0201,
|
|
"num_input_tokens_seen": 18278544,
|
|
"step": 735
|
|
},
|
|
{
|
|
"epoch": 17.159420289855074,
|
|
"grad_norm": 0.8654409050941467,
|
|
"learning_rate": 2.5652250036821523e-05,
|
|
"loss": 0.017,
|
|
"num_input_tokens_seen": 18396700,
|
|
"step": 740
|
|
},
|
|
{
|
|
"epoch": 17.27536231884058,
|
|
"grad_norm": 0.9699842929840088,
|
|
"learning_rate": 2.5391378445100644e-05,
|
|
"loss": 0.0187,
|
|
"num_input_tokens_seen": 18506229,
|
|
"step": 745
|
|
},
|
|
{
|
|
"epoch": 17.391304347826086,
|
|
"grad_norm": 0.8799194693565369,
|
|
"learning_rate": 2.5130464219022992e-05,
|
|
"loss": 0.0242,
|
|
"num_input_tokens_seen": 18621580,
|
|
"step": 750
|
|
},
|
|
{
|
|
"epoch": 17.507246376811594,
|
|
"grad_norm": 0.9715821146965027,
|
|
"learning_rate": 2.486953578097702e-05,
|
|
"loss": 0.0153,
|
|
"num_input_tokens_seen": 18748382,
|
|
"step": 755
|
|
},
|
|
{
|
|
"epoch": 17.6231884057971,
|
|
"grad_norm": 0.8819458484649658,
|
|
"learning_rate": 2.4608621554899362e-05,
|
|
"loss": 0.0182,
|
|
"num_input_tokens_seen": 18884730,
|
|
"step": 760
|
|
},
|
|
{
|
|
"epoch": 17.73913043478261,
|
|
"grad_norm": 0.8835431933403015,
|
|
"learning_rate": 2.4347749963178486e-05,
|
|
"loss": 0.0143,
|
|
"num_input_tokens_seen": 19003589,
|
|
"step": 765
|
|
},
|
|
{
|
|
"epoch": 17.855072463768117,
|
|
"grad_norm": 0.780754566192627,
|
|
"learning_rate": 2.4086949423558526e-05,
|
|
"loss": 0.0164,
|
|
"num_input_tokens_seen": 19136411,
|
|
"step": 770
|
|
},
|
|
{
|
|
"epoch": 17.971014492753625,
|
|
"grad_norm": 0.7591371536254883,
|
|
"learning_rate": 2.3826248346043663e-05,
|
|
"loss": 0.0157,
|
|
"num_input_tokens_seen": 19260436,
|
|
"step": 775
|
|
},
|
|
{
|
|
"epoch": 18.08695652173913,
|
|
"grad_norm": 0.673797070980072,
|
|
"learning_rate": 2.356567512980326e-05,
|
|
"loss": 0.0304,
|
|
"num_input_tokens_seen": 19388733,
|
|
"step": 780
|
|
},
|
|
{
|
|
"epoch": 18.202898550724637,
|
|
"grad_norm": 0.4008718729019165,
|
|
"learning_rate": 2.3305258160078274e-05,
|
|
"loss": 0.009,
|
|
"num_input_tokens_seen": 19531204,
|
|
"step": 785
|
|
},
|
|
{
|
|
"epoch": 18.318840579710145,
|
|
"grad_norm": 0.6676005125045776,
|
|
"learning_rate": 2.3045025805089118e-05,
|
|
"loss": 0.0105,
|
|
"num_input_tokens_seen": 19624608,
|
|
"step": 790
|
|
},
|
|
{
|
|
"epoch": 18.434782608695652,
|
|
"grad_norm": 0.6956990957260132,
|
|
"learning_rate": 2.278500641294543e-05,
|
|
"loss": 0.0104,
|
|
"num_input_tokens_seen": 19751062,
|
|
"step": 795
|
|
},
|
|
{
|
|
"epoch": 18.55072463768116,
|
|
"grad_norm": 0.80479896068573,
|
|
"learning_rate": 2.252522830855798e-05,
|
|
"loss": 0.0103,
|
|
"num_input_tokens_seen": 19879837,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 18.666666666666668,
|
|
"grad_norm": 0.7206840515136719,
|
|
"learning_rate": 2.2265719790553147e-05,
|
|
"loss": 0.0107,
|
|
"num_input_tokens_seen": 20019385,
|
|
"step": 805
|
|
},
|
|
{
|
|
"epoch": 18.782608695652176,
|
|
"grad_norm": 0.6994977593421936,
|
|
"learning_rate": 2.2006509128190195e-05,
|
|
"loss": 0.0269,
|
|
"num_input_tokens_seen": 20138003,
|
|
"step": 810
|
|
},
|
|
{
|
|
"epoch": 18.89855072463768,
|
|
"grad_norm": 0.5642988681793213,
|
|
"learning_rate": 2.174762455828187e-05,
|
|
"loss": 0.0086,
|
|
"num_input_tokens_seen": 20260523,
|
|
"step": 815
|
|
},
|
|
{
|
|
"epoch": 19.014492753623188,
|
|
"grad_norm": 0.5547834038734436,
|
|
"learning_rate": 2.1489094282118395e-05,
|
|
"loss": 0.0133,
|
|
"num_input_tokens_seen": 20375322,
|
|
"step": 820
|
|
},
|
|
{
|
|
"epoch": 19.130434782608695,
|
|
"grad_norm": 0.48678871989250183,
|
|
"learning_rate": 2.123094646239541e-05,
|
|
"loss": 0.0114,
|
|
"num_input_tokens_seen": 20477407,
|
|
"step": 825
|
|
},
|
|
{
|
|
"epoch": 19.246376811594203,
|
|
"grad_norm": 0.4791460633277893,
|
|
"learning_rate": 2.0973209220146135e-05,
|
|
"loss": 0.007,
|
|
"num_input_tokens_seen": 20605728,
|
|
"step": 830
|
|
},
|
|
{
|
|
"epoch": 19.36231884057971,
|
|
"grad_norm": 1.1198338270187378,
|
|
"learning_rate": 2.0715910631677968e-05,
|
|
"loss": 0.0088,
|
|
"num_input_tokens_seen": 20725799,
|
|
"step": 835
|
|
},
|
|
{
|
|
"epoch": 19.47826086956522,
|
|
"grad_norm": 0.6645247936248779,
|
|
"learning_rate": 2.0459078725514092e-05,
|
|
"loss": 0.007,
|
|
"num_input_tokens_seen": 20865534,
|
|
"step": 840
|
|
},
|
|
{
|
|
"epoch": 19.594202898550726,
|
|
"grad_norm": 0.5324479341506958,
|
|
"learning_rate": 2.020274147934019e-05,
|
|
"loss": 0.0059,
|
|
"num_input_tokens_seen": 20977913,
|
|
"step": 845
|
|
},
|
|
{
|
|
"epoch": 19.71014492753623,
|
|
"grad_norm": 0.6183504462242126,
|
|
"learning_rate": 1.9946926816956743e-05,
|
|
"loss": 0.0069,
|
|
"num_input_tokens_seen": 21102848,
|
|
"step": 850
|
|
},
|
|
{
|
|
"epoch": 19.82608695652174,
|
|
"grad_norm": 0.6665703058242798,
|
|
"learning_rate": 1.9691662605237166e-05,
|
|
"loss": 0.008,
|
|
"num_input_tokens_seen": 21243679,
|
|
"step": 855
|
|
},
|
|
{
|
|
"epoch": 19.942028985507246,
|
|
"grad_norm": 0.3298584222793579,
|
|
"learning_rate": 1.9436976651092144e-05,
|
|
"loss": 0.0127,
|
|
"num_input_tokens_seen": 21364202,
|
|
"step": 860
|
|
},
|
|
{
|
|
"epoch": 20.057971014492754,
|
|
"grad_norm": 0.2818591296672821,
|
|
"learning_rate": 1.9182896698440584e-05,
|
|
"loss": 0.0059,
|
|
"num_input_tokens_seen": 21496089,
|
|
"step": 865
|
|
},
|
|
{
|
|
"epoch": 20.17391304347826,
|
|
"grad_norm": 0.6906440258026123,
|
|
"learning_rate": 1.89294504251873e-05,
|
|
"loss": 0.0046,
|
|
"num_input_tokens_seen": 21603193,
|
|
"step": 870
|
|
},
|
|
{
|
|
"epoch": 20.28985507246377,
|
|
"grad_norm": 0.33482542634010315,
|
|
"learning_rate": 1.867666544020798e-05,
|
|
"loss": 0.0058,
|
|
"num_input_tokens_seen": 21742062,
|
|
"step": 875
|
|
},
|
|
{
|
|
"epoch": 20.405797101449274,
|
|
"grad_norm": 2.443847417831421,
|
|
"learning_rate": 1.8424569280341653e-05,
|
|
"loss": 0.0082,
|
|
"num_input_tokens_seen": 21869307,
|
|
"step": 880
|
|
},
|
|
{
|
|
"epoch": 20.52173913043478,
|
|
"grad_norm": 0.43886587023735046,
|
|
"learning_rate": 1.817318940739098e-05,
|
|
"loss": 0.0148,
|
|
"num_input_tokens_seen": 21992573,
|
|
"step": 885
|
|
},
|
|
{
|
|
"epoch": 20.63768115942029,
|
|
"grad_norm": 0.93570876121521,
|
|
"learning_rate": 1.7922553205130707e-05,
|
|
"loss": 0.0064,
|
|
"num_input_tokens_seen": 22101845,
|
|
"step": 890
|
|
},
|
|
{
|
|
"epoch": 20.753623188405797,
|
|
"grad_norm": 1176.9595947265625,
|
|
"learning_rate": 1.767268797632472e-05,
|
|
"loss": 0.008,
|
|
"num_input_tokens_seen": 22230253,
|
|
"step": 895
|
|
},
|
|
{
|
|
"epoch": 20.869565217391305,
|
|
"grad_norm": 0.35642215609550476,
|
|
"learning_rate": 1.7423620939751788e-05,
|
|
"loss": 0.0053,
|
|
"num_input_tokens_seen": 22373454,
|
|
"step": 900
|
|
},
|
|
{
|
|
"epoch": 20.985507246376812,
|
|
"grad_norm": 0.39736178517341614,
|
|
"learning_rate": 1.7175379227240523e-05,
|
|
"loss": 0.0054,
|
|
"num_input_tokens_seen": 22493123,
|
|
"step": 905
|
|
},
|
|
{
|
|
"epoch": 21.10144927536232,
|
|
"grad_norm": 0.5092463493347168,
|
|
"learning_rate": 1.692798988071385e-05,
|
|
"loss": 0.0044,
|
|
"num_input_tokens_seen": 22629005,
|
|
"step": 910
|
|
},
|
|
{
|
|
"epoch": 21.217391304347824,
|
|
"grad_norm": 0.26361697912216187,
|
|
"learning_rate": 1.6681479849243153e-05,
|
|
"loss": 0.0043,
|
|
"num_input_tokens_seen": 22752358,
|
|
"step": 915
|
|
},
|
|
{
|
|
"epoch": 21.333333333333332,
|
|
"grad_norm": 0.19933666288852692,
|
|
"learning_rate": 1.6435875986112685e-05,
|
|
"loss": 0.0035,
|
|
"num_input_tokens_seen": 22880349,
|
|
"step": 920
|
|
},
|
|
{
|
|
"epoch": 21.44927536231884,
|
|
"grad_norm": 0.22622954845428467,
|
|
"learning_rate": 1.6191205045894283e-05,
|
|
"loss": 0.0044,
|
|
"num_input_tokens_seen": 22987343,
|
|
"step": 925
|
|
},
|
|
{
|
|
"epoch": 21.565217391304348,
|
|
"grad_norm": 0.30199098587036133,
|
|
"learning_rate": 1.594749368153292e-05,
|
|
"loss": 0.0178,
|
|
"num_input_tokens_seen": 23113462,
|
|
"step": 930
|
|
},
|
|
{
|
|
"epoch": 21.681159420289855,
|
|
"grad_norm": 0.9627483487129211,
|
|
"learning_rate": 1.570476844144329e-05,
|
|
"loss": 0.0089,
|
|
"num_input_tokens_seen": 23221714,
|
|
"step": 935
|
|
},
|
|
{
|
|
"epoch": 21.797101449275363,
|
|
"grad_norm": 0.27791452407836914,
|
|
"learning_rate": 1.546305576661776e-05,
|
|
"loss": 0.004,
|
|
"num_input_tokens_seen": 23368857,
|
|
"step": 940
|
|
},
|
|
{
|
|
"epoch": 21.91304347826087,
|
|
"grad_norm": 0.3269965648651123,
|
|
"learning_rate": 1.5222381987746104e-05,
|
|
"loss": 0.004,
|
|
"num_input_tokens_seen": 23494483,
|
|
"step": 945
|
|
},
|
|
{
|
|
"epoch": 22.028985507246375,
|
|
"grad_norm": 0.15966826677322388,
|
|
"learning_rate": 1.4982773322347144e-05,
|
|
"loss": 0.0034,
|
|
"num_input_tokens_seen": 23605463,
|
|
"step": 950
|
|
},
|
|
{
|
|
"epoch": 22.144927536231883,
|
|
"grad_norm": 0.3009255826473236,
|
|
"learning_rate": 1.4744255871912823e-05,
|
|
"loss": 0.0066,
|
|
"num_input_tokens_seen": 23715776,
|
|
"step": 955
|
|
},
|
|
{
|
|
"epoch": 22.26086956521739,
|
|
"grad_norm": 0.4215935170650482,
|
|
"learning_rate": 1.4506855619064846e-05,
|
|
"loss": 0.0034,
|
|
"num_input_tokens_seen": 23841669,
|
|
"step": 960
|
|
},
|
|
{
|
|
"epoch": 22.3768115942029,
|
|
"grad_norm": 0.20214155316352844,
|
|
"learning_rate": 1.4270598424724292e-05,
|
|
"loss": 0.0032,
|
|
"num_input_tokens_seen": 23960567,
|
|
"step": 965
|
|
},
|
|
{
|
|
"epoch": 22.492753623188406,
|
|
"grad_norm": 7.0683207511901855,
|
|
"learning_rate": 1.4035510025294462e-05,
|
|
"loss": 0.0124,
|
|
"num_input_tokens_seen": 24074628,
|
|
"step": 970
|
|
},
|
|
{
|
|
"epoch": 22.608695652173914,
|
|
"grad_norm": 0.20178793370723724,
|
|
"learning_rate": 1.3801616029857378e-05,
|
|
"loss": 0.0027,
|
|
"num_input_tokens_seen": 24214324,
|
|
"step": 975
|
|
},
|
|
{
|
|
"epoch": 22.72463768115942,
|
|
"grad_norm": 1.3855236768722534,
|
|
"learning_rate": 1.3568941917384036e-05,
|
|
"loss": 0.0037,
|
|
"num_input_tokens_seen": 24326727,
|
|
"step": 980
|
|
},
|
|
{
|
|
"epoch": 22.840579710144926,
|
|
"grad_norm": 0.18420317769050598,
|
|
"learning_rate": 1.3337513033958904e-05,
|
|
"loss": 0.0029,
|
|
"num_input_tokens_seen": 24456961,
|
|
"step": 985
|
|
},
|
|
{
|
|
"epoch": 22.956521739130434,
|
|
"grad_norm": 0.15907694399356842,
|
|
"learning_rate": 1.310735459001884e-05,
|
|
"loss": 0.0035,
|
|
"num_input_tokens_seen": 24606652,
|
|
"step": 990
|
|
},
|
|
{
|
|
"epoch": 23.07246376811594,
|
|
"grad_norm": 0.2548115849494934,
|
|
"learning_rate": 1.2878491657606872e-05,
|
|
"loss": 0.002,
|
|
"num_input_tokens_seen": 24710410,
|
|
"step": 995
|
|
},
|
|
{
|
|
"epoch": 23.18840579710145,
|
|
"grad_norm": 0.36587971448898315,
|
|
"learning_rate": 1.2650949167640993e-05,
|
|
"loss": 0.0023,
|
|
"num_input_tokens_seen": 24831908,
|
|
"step": 1000
|
|
},
|
|
{
|
|
"epoch": 23.304347826086957,
|
|
"grad_norm": 0.13662408292293549,
|
|
"learning_rate": 1.2424751907198312e-05,
|
|
"loss": 0.0031,
|
|
"num_input_tokens_seen": 24951342,
|
|
"step": 1005
|
|
},
|
|
{
|
|
"epoch": 23.420289855072465,
|
|
"grad_norm": 0.19979843497276306,
|
|
"learning_rate": 1.2199924516814939e-05,
|
|
"loss": 0.0027,
|
|
"num_input_tokens_seen": 25088309,
|
|
"step": 1010
|
|
},
|
|
{
|
|
"epoch": 23.536231884057973,
|
|
"grad_norm": 0.14170995354652405,
|
|
"learning_rate": 1.1976491487801748e-05,
|
|
"loss": 0.0124,
|
|
"num_input_tokens_seen": 25216080,
|
|
"step": 1015
|
|
},
|
|
{
|
|
"epoch": 23.652173913043477,
|
|
"grad_norm": 0.06863216310739517,
|
|
"learning_rate": 1.1754477159576499e-05,
|
|
"loss": 0.0023,
|
|
"num_input_tokens_seen": 25326581,
|
|
"step": 1020
|
|
},
|
|
{
|
|
"epoch": 23.768115942028984,
|
|
"grad_norm": 0.25133436918258667,
|
|
"learning_rate": 1.1533905717012428e-05,
|
|
"loss": 0.0027,
|
|
"num_input_tokens_seen": 25477500,
|
|
"step": 1025
|
|
},
|
|
{
|
|
"epoch": 23.884057971014492,
|
|
"grad_norm": 0.28348398208618164,
|
|
"learning_rate": 1.1314801187803686e-05,
|
|
"loss": 0.0041,
|
|
"num_input_tokens_seen": 25601354,
|
|
"step": 1030
|
|
},
|
|
{
|
|
"epoch": 24.0,
|
|
"grad_norm": 0.5024954676628113,
|
|
"learning_rate": 1.1097187439847939e-05,
|
|
"loss": 0.0021,
|
|
"num_input_tokens_seen": 25719240,
|
|
"step": 1035
|
|
},
|
|
{
|
|
"epoch": 24.115942028985508,
|
|
"grad_norm": 0.1774568408727646,
|
|
"learning_rate": 1.088108817864629e-05,
|
|
"loss": 0.0039,
|
|
"num_input_tokens_seen": 25834910,
|
|
"step": 1040
|
|
},
|
|
{
|
|
"epoch": 24.231884057971016,
|
|
"grad_norm": 0.08105342090129852,
|
|
"learning_rate": 1.0666526944721016e-05,
|
|
"loss": 0.0025,
|
|
"num_input_tokens_seen": 25974530,
|
|
"step": 1045
|
|
},
|
|
{
|
|
"epoch": 24.347826086956523,
|
|
"grad_norm": 0.13048779964447021,
|
|
"learning_rate": 1.0453527111051184e-05,
|
|
"loss": 0.002,
|
|
"num_input_tokens_seen": 26104464,
|
|
"step": 1050
|
|
},
|
|
{
|
|
"epoch": 24.463768115942027,
|
|
"grad_norm": 0.10774020105600357,
|
|
"learning_rate": 1.0242111880526495e-05,
|
|
"loss": 0.0024,
|
|
"num_input_tokens_seen": 26251334,
|
|
"step": 1055
|
|
},
|
|
{
|
|
"epoch": 24.579710144927535,
|
|
"grad_norm": 0.7494776248931885,
|
|
"learning_rate": 1.003230428341979e-05,
|
|
"loss": 0.0031,
|
|
"num_input_tokens_seen": 26366561,
|
|
"step": 1060
|
|
},
|
|
{
|
|
"epoch": 24.695652173913043,
|
|
"grad_norm": 0.3580308258533478,
|
|
"learning_rate": 9.824127174878195e-06,
|
|
"loss": 0.0022,
|
|
"num_input_tokens_seen": 26486437,
|
|
"step": 1065
|
|
},
|
|
{
|
|
"epoch": 24.81159420289855,
|
|
"grad_norm": 0.1473228931427002,
|
|
"learning_rate": 9.617603232433475e-06,
|
|
"loss": 0.0022,
|
|
"num_input_tokens_seen": 26601526,
|
|
"step": 1070
|
|
},
|
|
{
|
|
"epoch": 24.92753623188406,
|
|
"grad_norm": 0.11716706305742264,
|
|
"learning_rate": 9.412754953531663e-06,
|
|
"loss": 0.0109,
|
|
"num_input_tokens_seen": 26727922,
|
|
"step": 1075
|
|
},
|
|
{
|
|
"epoch": 25.043478260869566,
|
|
"grad_norm": 0.12043190747499466,
|
|
"learning_rate": 9.209604653082326e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 26835621,
|
|
"step": 1080
|
|
},
|
|
{
|
|
"epoch": 25.159420289855074,
|
|
"grad_norm": 0.1277165412902832,
|
|
"learning_rate": 9.008174461027724e-06,
|
|
"loss": 0.0016,
|
|
"num_input_tokens_seen": 26955101,
|
|
"step": 1085
|
|
},
|
|
{
|
|
"epoch": 25.27536231884058,
|
|
"grad_norm": 0.08892516791820526,
|
|
"learning_rate": 8.808486319932083e-06,
|
|
"loss": 0.002,
|
|
"num_input_tokens_seen": 27077833,
|
|
"step": 1090
|
|
},
|
|
{
|
|
"epoch": 25.391304347826086,
|
|
"grad_norm": 0.30754807591438293,
|
|
"learning_rate": 8.610561982591357e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 27192758,
|
|
"step": 1095
|
|
},
|
|
{
|
|
"epoch": 25.507246376811594,
|
|
"grad_norm": 0.7194050550460815,
|
|
"learning_rate": 8.414423009663563e-06,
|
|
"loss": 0.0028,
|
|
"num_input_tokens_seen": 27324970,
|
|
"step": 1100
|
|
},
|
|
{
|
|
"epoch": 25.6231884057971,
|
|
"grad_norm": 4777.61328125,
|
|
"learning_rate": 8.220090767320137e-06,
|
|
"loss": 0.0021,
|
|
"num_input_tokens_seen": 27477531,
|
|
"step": 1105
|
|
},
|
|
{
|
|
"epoch": 25.73913043478261,
|
|
"grad_norm": 2.280327081680298,
|
|
"learning_rate": 8.027586424918412e-06,
|
|
"loss": 0.0057,
|
|
"num_input_tokens_seen": 27592035,
|
|
"step": 1110
|
|
},
|
|
{
|
|
"epoch": 25.855072463768117,
|
|
"grad_norm": 0.13882993161678314,
|
|
"learning_rate": 7.836930952695533e-06,
|
|
"loss": 0.0067,
|
|
"num_input_tokens_seen": 27712377,
|
|
"step": 1115
|
|
},
|
|
{
|
|
"epoch": 25.971014492753625,
|
|
"grad_norm": 0.20987676084041595,
|
|
"learning_rate": 7.648145119484153e-06,
|
|
"loss": 0.002,
|
|
"num_input_tokens_seen": 27834613,
|
|
"step": 1120
|
|
},
|
|
{
|
|
"epoch": 26.08695652173913,
|
|
"grad_norm": 0.09795770049095154,
|
|
"learning_rate": 7.461249490449954e-06,
|
|
"loss": 0.0021,
|
|
"num_input_tokens_seen": 27966996,
|
|
"step": 1125
|
|
},
|
|
{
|
|
"epoch": 26.202898550724637,
|
|
"grad_norm": 0.14506971836090088,
|
|
"learning_rate": 7.276264424851423e-06,
|
|
"loss": 0.002,
|
|
"num_input_tokens_seen": 28093538,
|
|
"step": 1130
|
|
},
|
|
{
|
|
"epoch": 26.318840579710145,
|
|
"grad_norm": 0.08091314136981964,
|
|
"learning_rate": 7.0932100738220265e-06,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 28215579,
|
|
"step": 1135
|
|
},
|
|
{
|
|
"epoch": 26.434782608695652,
|
|
"grad_norm": 0.22550061345100403,
|
|
"learning_rate": 6.912106378175098e-06,
|
|
"loss": 0.0014,
|
|
"num_input_tokens_seen": 28344144,
|
|
"step": 1140
|
|
},
|
|
{
|
|
"epoch": 26.55072463768116,
|
|
"grad_norm": 0.23987355828285217,
|
|
"learning_rate": 6.732973066231563e-06,
|
|
"loss": 0.0022,
|
|
"num_input_tokens_seen": 28478650,
|
|
"step": 1145
|
|
},
|
|
{
|
|
"epoch": 26.666666666666668,
|
|
"grad_norm": 0.1993756741285324,
|
|
"learning_rate": 6.555829651670911e-06,
|
|
"loss": 0.0023,
|
|
"num_input_tokens_seen": 28593004,
|
|
"step": 1150
|
|
},
|
|
{
|
|
"epoch": 26.782608695652176,
|
|
"grad_norm": 0.7184757590293884,
|
|
"learning_rate": 6.380695431405456e-06,
|
|
"loss": 0.0028,
|
|
"num_input_tokens_seen": 28707392,
|
|
"step": 1155
|
|
},
|
|
{
|
|
"epoch": 26.89855072463768,
|
|
"grad_norm": 0.06247011199593544,
|
|
"learning_rate": 6.207589483478266e-06,
|
|
"loss": 0.006,
|
|
"num_input_tokens_seen": 28834902,
|
|
"step": 1160
|
|
},
|
|
{
|
|
"epoch": 27.014492753623188,
|
|
"grad_norm": 0.11046591401100159,
|
|
"learning_rate": 6.0365306649849214e-06,
|
|
"loss": 0.0045,
|
|
"num_input_tokens_seen": 28948812,
|
|
"step": 1165
|
|
},
|
|
{
|
|
"epoch": 27.130434782608695,
|
|
"grad_norm": 0.12309098988771439,
|
|
"learning_rate": 5.867537610019317e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 29078309,
|
|
"step": 1170
|
|
},
|
|
{
|
|
"epoch": 27.246376811594203,
|
|
"grad_norm": 0.11428932845592499,
|
|
"learning_rate": 5.700628727643806e-06,
|
|
"loss": 0.002,
|
|
"num_input_tokens_seen": 29211503,
|
|
"step": 1175
|
|
},
|
|
{
|
|
"epoch": 27.36231884057971,
|
|
"grad_norm": 0.1093268170952797,
|
|
"learning_rate": 5.53582219988382e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 29344489,
|
|
"step": 1180
|
|
},
|
|
{
|
|
"epoch": 27.47826086956522,
|
|
"grad_norm": 0.2166384607553482,
|
|
"learning_rate": 5.373135979747227e-06,
|
|
"loss": 0.006,
|
|
"num_input_tokens_seen": 29464082,
|
|
"step": 1185
|
|
},
|
|
{
|
|
"epoch": 27.594202898550726,
|
|
"grad_norm": 0.15387850999832153,
|
|
"learning_rate": 5.2125877892686496e-06,
|
|
"loss": 0.0043,
|
|
"num_input_tokens_seen": 29581124,
|
|
"step": 1190
|
|
},
|
|
{
|
|
"epoch": 27.71014492753623,
|
|
"grad_norm": 0.11962082982063293,
|
|
"learning_rate": 5.054195117578914e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 29696346,
|
|
"step": 1195
|
|
},
|
|
{
|
|
"epoch": 27.82608695652174,
|
|
"grad_norm": 0.18724732100963593,
|
|
"learning_rate": 4.897975218999926e-06,
|
|
"loss": 0.002,
|
|
"num_input_tokens_seen": 29815117,
|
|
"step": 1200
|
|
},
|
|
{
|
|
"epoch": 27.942028985507246,
|
|
"grad_norm": 0.09917350113391876,
|
|
"learning_rate": 4.743945111165068e-06,
|
|
"loss": 0.0022,
|
|
"num_input_tokens_seen": 29939175,
|
|
"step": 1205
|
|
},
|
|
{
|
|
"epoch": 28.057971014492754,
|
|
"grad_norm": 0.08235369622707367,
|
|
"learning_rate": 4.592121573165414e-06,
|
|
"loss": 0.0016,
|
|
"num_input_tokens_seen": 30079840,
|
|
"step": 1210
|
|
},
|
|
{
|
|
"epoch": 28.17391304347826,
|
|
"grad_norm": 0.20488996803760529,
|
|
"learning_rate": 4.442521143721892e-06,
|
|
"loss": 0.0033,
|
|
"num_input_tokens_seen": 30192219,
|
|
"step": 1215
|
|
},
|
|
{
|
|
"epoch": 28.28985507246377,
|
|
"grad_norm": 0.05383768677711487,
|
|
"learning_rate": 4.295160119383712e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 30330969,
|
|
"step": 1220
|
|
},
|
|
{
|
|
"epoch": 28.405797101449274,
|
|
"grad_norm": 0.14237363636493683,
|
|
"learning_rate": 4.150054552753055e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 30453302,
|
|
"step": 1225
|
|
},
|
|
{
|
|
"epoch": 28.52173913043478,
|
|
"grad_norm": 0.12487669289112091,
|
|
"learning_rate": 4.007220250736454e-06,
|
|
"loss": 0.0078,
|
|
"num_input_tokens_seen": 30568943,
|
|
"step": 1230
|
|
},
|
|
{
|
|
"epoch": 28.63768115942029,
|
|
"grad_norm": 0.1423855572938919,
|
|
"learning_rate": 3.866672772822863e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 30696057,
|
|
"step": 1235
|
|
},
|
|
{
|
|
"epoch": 28.753623188405797,
|
|
"grad_norm": 0.1543101817369461,
|
|
"learning_rate": 3.7284274293887115e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 30815506,
|
|
"step": 1240
|
|
},
|
|
{
|
|
"epoch": 28.869565217391305,
|
|
"grad_norm": 0.1402539610862732,
|
|
"learning_rate": 3.592499280030057e-06,
|
|
"loss": 0.0027,
|
|
"num_input_tokens_seen": 30916446,
|
|
"step": 1245
|
|
},
|
|
{
|
|
"epoch": 28.985507246376812,
|
|
"grad_norm": 0.26191645860671997,
|
|
"learning_rate": 3.458903131922134e-06,
|
|
"loss": 0.0023,
|
|
"num_input_tokens_seen": 31054242,
|
|
"step": 1250
|
|
},
|
|
{
|
|
"epoch": 29.10144927536232,
|
|
"grad_norm": 0.09874732792377472,
|
|
"learning_rate": 3.3276535382063213e-06,
|
|
"loss": 0.0029,
|
|
"num_input_tokens_seen": 31189078,
|
|
"step": 1255
|
|
},
|
|
{
|
|
"epoch": 29.217391304347824,
|
|
"grad_norm": 0.11677820980548859,
|
|
"learning_rate": 3.198764796404807e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 31311374,
|
|
"step": 1260
|
|
},
|
|
{
|
|
"epoch": 29.333333333333332,
|
|
"grad_norm": 0.05459802597761154,
|
|
"learning_rate": 3.0722509468631392e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 31444681,
|
|
"step": 1265
|
|
},
|
|
{
|
|
"epoch": 29.44927536231884,
|
|
"grad_norm": 0.1113714948296547,
|
|
"learning_rate": 2.948125771220697e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 31567569,
|
|
"step": 1270
|
|
},
|
|
{
|
|
"epoch": 29.565217391304348,
|
|
"grad_norm": 0.1816156655550003,
|
|
"learning_rate": 2.8264027909094715e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 31697338,
|
|
"step": 1275
|
|
},
|
|
{
|
|
"epoch": 29.681159420289855,
|
|
"grad_norm": 0.13639949262142181,
|
|
"learning_rate": 2.707095265681081e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 31826661,
|
|
"step": 1280
|
|
},
|
|
{
|
|
"epoch": 29.797101449275363,
|
|
"grad_norm": 0.05292365327477455,
|
|
"learning_rate": 2.5902161921623454e-06,
|
|
"loss": 0.0023,
|
|
"num_input_tokens_seen": 31944680,
|
|
"step": 1285
|
|
},
|
|
{
|
|
"epoch": 29.91304347826087,
|
|
"grad_norm": 0.16608740389347076,
|
|
"learning_rate": 2.475778302439524e-06,
|
|
"loss": 0.0078,
|
|
"num_input_tokens_seen": 32067106,
|
|
"step": 1290
|
|
},
|
|
{
|
|
"epoch": 30.028985507246375,
|
|
"grad_norm": 0.09277443587779999,
|
|
"learning_rate": 2.3637940626713346e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 32184526,
|
|
"step": 1295
|
|
},
|
|
{
|
|
"epoch": 30.144927536231883,
|
|
"grad_norm": 0.18832191824913025,
|
|
"learning_rate": 2.254275671731007e-06,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 32309423,
|
|
"step": 1300
|
|
},
|
|
{
|
|
"epoch": 30.26086956521739,
|
|
"grad_norm": 0.1828456073999405,
|
|
"learning_rate": 2.14723505987737e-06,
|
|
"loss": 0.0071,
|
|
"num_input_tokens_seen": 32429445,
|
|
"step": 1305
|
|
},
|
|
{
|
|
"epoch": 30.3768115942029,
|
|
"grad_norm": 0.07503814995288849,
|
|
"learning_rate": 2.0426838874552714e-06,
|
|
"loss": 0.0016,
|
|
"num_input_tokens_seen": 32540571,
|
|
"step": 1310
|
|
},
|
|
{
|
|
"epoch": 30.492753623188406,
|
|
"grad_norm": 0.19047732651233673,
|
|
"learning_rate": 1.9406335436253724e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 32665528,
|
|
"step": 1315
|
|
},
|
|
{
|
|
"epoch": 30.608695652173914,
|
|
"grad_norm": 0.17791509628295898,
|
|
"learning_rate": 1.8410951451234533e-06,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 32800773,
|
|
"step": 1320
|
|
},
|
|
{
|
|
"epoch": 30.72463768115942,
|
|
"grad_norm": 0.10698456317186356,
|
|
"learning_rate": 1.7440795350494588e-06,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 32928397,
|
|
"step": 1325
|
|
},
|
|
{
|
|
"epoch": 30.840579710144926,
|
|
"grad_norm": 0.0963551327586174,
|
|
"learning_rate": 1.649597281686302e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 33054819,
|
|
"step": 1330
|
|
},
|
|
{
|
|
"epoch": 30.956521739130434,
|
|
"grad_norm": 0.24703514575958252,
|
|
"learning_rate": 1.5576586773486195e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 33180616,
|
|
"step": 1335
|
|
},
|
|
{
|
|
"epoch": 31.07246376811594,
|
|
"grad_norm": 0.12497910857200623,
|
|
"learning_rate": 1.4682737372615967e-06,
|
|
"loss": 0.0038,
|
|
"num_input_tokens_seen": 33298041,
|
|
"step": 1340
|
|
},
|
|
{
|
|
"epoch": 31.18840579710145,
|
|
"grad_norm": 0.18260960280895233,
|
|
"learning_rate": 1.3814521984699596e-06,
|
|
"loss": 0.0052,
|
|
"num_input_tokens_seen": 33408343,
|
|
"step": 1345
|
|
},
|
|
{
|
|
"epoch": 31.304347826086957,
|
|
"grad_norm": 0.13422255218029022,
|
|
"learning_rate": 1.297203518777293e-06,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 33545364,
|
|
"step": 1350
|
|
},
|
|
{
|
|
"epoch": 31.420289855072465,
|
|
"grad_norm": 0.1285027116537094,
|
|
"learning_rate": 1.2155368757157643e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 33652900,
|
|
"step": 1355
|
|
},
|
|
{
|
|
"epoch": 31.536231884057973,
|
|
"grad_norm": 0.12832242250442505,
|
|
"learning_rate": 1.1364611655463736e-06,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 33768791,
|
|
"step": 1360
|
|
},
|
|
{
|
|
"epoch": 31.652173913043477,
|
|
"grad_norm": 0.12093157321214676,
|
|
"learning_rate": 1.0599850022898539e-06,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 33892837,
|
|
"step": 1365
|
|
},
|
|
{
|
|
"epoch": 31.768115942028984,
|
|
"grad_norm": 0.7227018475532532,
|
|
"learning_rate": 9.861167167883046e-07,
|
|
"loss": 0.0022,
|
|
"num_input_tokens_seen": 34015288,
|
|
"step": 1370
|
|
},
|
|
{
|
|
"epoch": 31.884057971014492,
|
|
"grad_norm": 2.143653631210327,
|
|
"learning_rate": 9.148643557976955e-07,
|
|
"loss": 0.0037,
|
|
"num_input_tokens_seen": 34154884,
|
|
"step": 1375
|
|
},
|
|
{
|
|
"epoch": 32.0,
|
|
"grad_norm": 0.17518474161624908,
|
|
"learning_rate": 8.462356811112987e-07,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 34292320,
|
|
"step": 1380
|
|
},
|
|
{
|
|
"epoch": 32.11594202898551,
|
|
"grad_norm": 0.1274159997701645,
|
|
"learning_rate": 7.802381687141535e-07,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 34413850,
|
|
"step": 1385
|
|
},
|
|
{
|
|
"epoch": 32.231884057971016,
|
|
"grad_norm": 0.11443401873111725,
|
|
"learning_rate": 7.168790079686932e-07,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 34547127,
|
|
"step": 1390
|
|
},
|
|
{
|
|
"epoch": 32.34782608695652,
|
|
"grad_norm": 0.08239752799272537,
|
|
"learning_rate": 6.561651008315738e-07,
|
|
"loss": 0.0035,
|
|
"num_input_tokens_seen": 34685112,
|
|
"step": 1395
|
|
},
|
|
{
|
|
"epoch": 32.46376811594203,
|
|
"grad_norm": 0.7361220717430115,
|
|
"learning_rate": 5.981030611018234e-07,
|
|
"loss": 0.0063,
|
|
"num_input_tokens_seen": 34810484,
|
|
"step": 1400
|
|
},
|
|
{
|
|
"epoch": 32.57971014492754,
|
|
"grad_norm": 0.20323431491851807,
|
|
"learning_rate": 5.426992137003622e-07,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 34920531,
|
|
"step": 1405
|
|
},
|
|
{
|
|
"epoch": 32.69565217391305,
|
|
"grad_norm": 0.11165229231119156,
|
|
"learning_rate": 4.899595939810236e-07,
|
|
"loss": 0.002,
|
|
"num_input_tokens_seen": 35035657,
|
|
"step": 1410
|
|
},
|
|
{
|
|
"epoch": 32.81159420289855,
|
|
"grad_norm": 0.15023387968540192,
|
|
"learning_rate": 4.398899470730827e-07,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 35167466,
|
|
"step": 1415
|
|
},
|
|
{
|
|
"epoch": 32.927536231884055,
|
|
"grad_norm": 0.18479810655117035,
|
|
"learning_rate": 3.9249572725543196e-07,
|
|
"loss": 0.0016,
|
|
"num_input_tokens_seen": 35296818,
|
|
"step": 1420
|
|
},
|
|
{
|
|
"epoch": 33.04347826086956,
|
|
"grad_norm": 0.08527754247188568,
|
|
"learning_rate": 3.477820973624063e-07,
|
|
"loss": 0.0015,
|
|
"num_input_tokens_seen": 35430399,
|
|
"step": 1425
|
|
},
|
|
{
|
|
"epoch": 33.15942028985507,
|
|
"grad_norm": 0.16888481378555298,
|
|
"learning_rate": 3.0575392822139726e-07,
|
|
"loss": 0.0057,
|
|
"num_input_tokens_seen": 35551540,
|
|
"step": 1430
|
|
},
|
|
{
|
|
"epoch": 33.27536231884058,
|
|
"grad_norm": 0.18187086284160614,
|
|
"learning_rate": 2.664157981222437e-07,
|
|
"loss": 0.0016,
|
|
"num_input_tokens_seen": 35676077,
|
|
"step": 1435
|
|
},
|
|
{
|
|
"epoch": 33.391304347826086,
|
|
"grad_norm": 0.15047162771224976,
|
|
"learning_rate": 2.297719923185032e-07,
|
|
"loss": 0.0016,
|
|
"num_input_tokens_seen": 35785127,
|
|
"step": 1440
|
|
},
|
|
{
|
|
"epoch": 33.507246376811594,
|
|
"grad_norm": 0.12288761883974075,
|
|
"learning_rate": 1.9582650256064205e-07,
|
|
"loss": 0.0019,
|
|
"num_input_tokens_seen": 35911682,
|
|
"step": 1445
|
|
},
|
|
{
|
|
"epoch": 33.6231884057971,
|
|
"grad_norm": 0.22509098052978516,
|
|
"learning_rate": 1.645830266611914e-07,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 36030754,
|
|
"step": 1450
|
|
},
|
|
{
|
|
"epoch": 33.73913043478261,
|
|
"grad_norm": 2.0408618450164795,
|
|
"learning_rate": 1.3604496809195288e-07,
|
|
"loss": 0.0042,
|
|
"num_input_tokens_seen": 36146749,
|
|
"step": 1455
|
|
},
|
|
{
|
|
"epoch": 33.85507246376812,
|
|
"grad_norm": 0.10705255717039108,
|
|
"learning_rate": 1.1021543561322012e-07,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 36278454,
|
|
"step": 1460
|
|
},
|
|
{
|
|
"epoch": 33.971014492753625,
|
|
"grad_norm": 1876.0384521484375,
|
|
"learning_rate": 8.709724293513854e-08,
|
|
"loss": 0.0017,
|
|
"num_input_tokens_seen": 36408834,
|
|
"step": 1465
|
|
},
|
|
{
|
|
"epoch": 34.08695652173913,
|
|
"grad_norm": 0.1927630454301834,
|
|
"learning_rate": 6.66929084112089e-08,
|
|
"loss": 0.0015,
|
|
"num_input_tokens_seen": 36550538,
|
|
"step": 1470
|
|
},
|
|
{
|
|
"epoch": 34.20289855072464,
|
|
"grad_norm": 0.1668202131986618,
|
|
"learning_rate": 4.900465476393168e-08,
|
|
"loss": 0.0018,
|
|
"num_input_tokens_seen": 36647436,
|
|
"step": 1475
|
|
},
|
|
{
|
|
"epoch": 34.31884057971015,
|
|
"grad_norm": 0.7123565673828125,
|
|
"learning_rate": 3.403440884269526e-08,
|
|
"loss": 0.0024,
|
|
"num_input_tokens_seen": 36785387,
|
|
"step": 1480
|
|
},
|
|
{
|
|
"epoch": 34.43478260869565,
|
|
"grad_norm": 0.16973845660686493,
|
|
"learning_rate": 2.1783801413866046e-08,
|
|
"loss": 0.0021,
|
|
"num_input_tokens_seen": 36915606,
|
|
"step": 1485
|
|
},
|
|
{
|
|
"epoch": 34.55072463768116,
|
|
"grad_norm": 2.034724473953247,
|
|
"learning_rate": 1.2254166983152737e-08,
|
|
"loss": 0.0035,
|
|
"num_input_tokens_seen": 37036117,
|
|
"step": 1490
|
|
},
|
|
{
|
|
"epoch": 34.666666666666664,
|
|
"grad_norm": 0.155415877699852,
|
|
"learning_rate": 5.446543650219904e-09,
|
|
"loss": 0.0016,
|
|
"num_input_tokens_seen": 37165587,
|
|
"step": 1495
|
|
},
|
|
{
|
|
"epoch": 34.78260869565217,
|
|
"grad_norm": 0.10199662297964096,
|
|
"learning_rate": 1.3616729956228425e-09,
|
|
"loss": 0.0015,
|
|
"num_input_tokens_seen": 37290827,
|
|
"step": 1500
|
|
},
|
|
{
|
|
"epoch": 34.89855072463768,
|
|
"grad_norm": 0.14740267395973206,
|
|
"learning_rate": 0.0,
|
|
"loss": 0.0053,
|
|
"num_input_tokens_seen": 37412688,
|
|
"step": 1505
|
|
},
|
|
{
|
|
"epoch": 34.89855072463768,
|
|
"num_input_tokens_seen": 37412688,
|
|
"step": 1505,
|
|
"total_flos": 8.033958240027034e+16,
|
|
"train_loss": 0.3889684765070578,
|
|
"train_runtime": 37510.9602,
|
|
"train_samples_per_second": 0.322,
|
|
"train_steps_per_second": 0.04
|
|
}
|
|
],
|
|
"logging_steps": 5,
|
|
"max_steps": 1505,
|
|
"num_input_tokens_seen": 37412688,
|
|
"num_train_epochs": 35,
|
|
"save_steps": 100,
|
|
"stateful_callbacks": {
|
|
"TrainerControl": {
|
|
"args": {
|
|
"should_epoch_stop": false,
|
|
"should_evaluate": false,
|
|
"should_log": false,
|
|
"should_save": true,
|
|
"should_training_stop": false
|
|
},
|
|
"attributes": {}
|
|
}
|
|
},
|
|
"total_flos": 8.033958240027034e+16,
|
|
"train_batch_size": 1,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|
|
|