{ "best_metric": 6.703342437744141, "best_model_checkpoint": "miner_id_24/checkpoint-50", "epoch": 0.0353198686697841, "eval_steps": 25, "global_step": 71, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0004974629390110437, "grad_norm": 2.0978946685791016, "learning_rate": 0.00015, "loss": 6.9589, "step": 1 }, { "epoch": 0.0004974629390110437, "eval_loss": 6.9487104415893555, "eval_runtime": 0.0398, "eval_samples_per_second": 1255.231, "eval_steps_per_second": 75.314, "step": 1 }, { "epoch": 0.0009949258780220873, "grad_norm": 0.784283459186554, "learning_rate": 0.0003, "loss": 6.9495, "step": 2 }, { "epoch": 0.0014923888170331311, "grad_norm": 1.2301733493804932, "learning_rate": 0.0002998600959423082, "loss": 6.9501, "step": 3 }, { "epoch": 0.0019898517560441747, "grad_norm": 1.3414157629013062, "learning_rate": 0.0002994406737417567, "loss": 6.947, "step": 4 }, { "epoch": 0.0024873146950552183, "grad_norm": 1.5272213220596313, "learning_rate": 0.00029874260271490463, "loss": 6.9409, "step": 5 }, { "epoch": 0.0029847776340662623, "grad_norm": 1.9261209964752197, "learning_rate": 0.00029776732972055516, "loss": 6.9326, "step": 6 }, { "epoch": 0.003482240573077306, "grad_norm": 1.9587503671646118, "learning_rate": 0.0002965168761609197, "loss": 6.9088, "step": 7 }, { "epoch": 0.003979703512088349, "grad_norm": 0.8569906949996948, "learning_rate": 0.0002949938337919529, "loss": 6.9137, "step": 8 }, { "epoch": 0.004477166451099393, "grad_norm": 1.1164758205413818, "learning_rate": 0.0002932013593515431, "loss": 6.9299, "step": 9 }, { "epoch": 0.0049746293901104365, "grad_norm": 1.2469490766525269, "learning_rate": 0.00029114316801669057, "loss": 6.8841, "step": 10 }, { "epoch": 0.0054720923291214805, "grad_norm": 1.5062528848648071, "learning_rate": 0.00028882352570323616, "loss": 6.8565, "step": 11 }, { "epoch": 0.0059695552681325245, "grad_norm": 1.8433338403701782, "learning_rate": 0.00028624724022409897, "loss": 6.822, "step": 12 }, { "epoch": 0.006467018207143568, "grad_norm": 1.1858328580856323, "learning_rate": 0.0002834196513243502, "loss": 6.8764, "step": 13 }, { "epoch": 0.006964481146154612, "grad_norm": 1.0784205198287964, "learning_rate": 0.0002803466196137759, "loss": 6.8753, "step": 14 }, { "epoch": 0.007461944085165655, "grad_norm": 0.946907639503479, "learning_rate": 0.00027703451441986836, "loss": 6.8735, "step": 15 }, { "epoch": 0.007959407024176699, "grad_norm": 1.3893108367919922, "learning_rate": 0.000273490200586422, "loss": 6.8587, "step": 16 }, { "epoch": 0.008456869963187742, "grad_norm": 0.709138810634613, "learning_rate": 0.00026972102424509665, "loss": 6.8263, "step": 17 }, { "epoch": 0.008954332902198787, "grad_norm": 1.2287291288375854, "learning_rate": 0.00026573479758943753, "loss": 6.7822, "step": 18 }, { "epoch": 0.00945179584120983, "grad_norm": 1.5594730377197266, "learning_rate": 0.0002615397826829114, "loss": 6.7248, "step": 19 }, { "epoch": 0.009949258780220873, "grad_norm": 1.0414823293685913, "learning_rate": 0.0002571446743345183, "loss": 6.8372, "step": 20 }, { "epoch": 0.010446721719231918, "grad_norm": 1.1876541376113892, "learning_rate": 0.00025255858207747205, "loss": 6.8816, "step": 21 }, { "epoch": 0.010944184658242961, "grad_norm": 1.0211689472198486, "learning_rate": 0.0002477910112883017, "loss": 6.836, "step": 22 }, { "epoch": 0.011441647597254004, "grad_norm": 1.1104573011398315, "learning_rate": 0.00024285184348550706, "loss": 6.7411, "step": 23 }, { "epoch": 0.011939110536265049, "grad_norm": 1.2712668180465698, "learning_rate": 0.0002377513158486027, "loss": 6.6863, "step": 24 }, { "epoch": 0.012436573475276092, "grad_norm": 1.0742238759994507, "learning_rate": 0.00023249999999999999, "loss": 6.7343, "step": 25 }, { "epoch": 0.012436573475276092, "eval_loss": 6.77644157409668, "eval_runtime": 0.0409, "eval_samples_per_second": 1222.387, "eval_steps_per_second": 73.343, "step": 25 }, { "epoch": 0.012934036414287135, "grad_norm": 1.5590488910675049, "learning_rate": 0.00022710878009370554, "loss": 6.8274, "step": 26 }, { "epoch": 0.013431499353298178, "grad_norm": 0.6883417367935181, "learning_rate": 0.00022158883025624965, "loss": 6.8454, "step": 27 }, { "epoch": 0.013928962292309223, "grad_norm": 1.1368870735168457, "learning_rate": 0.0002159515914266029, "loss": 6.8375, "step": 28 }, { "epoch": 0.014426425231320266, "grad_norm": 0.9605547785758972, "learning_rate": 0.0002102087476430831, "loss": 6.8131, "step": 29 }, { "epoch": 0.01492388817033131, "grad_norm": 1.0653886795043945, "learning_rate": 0.00020437220182640135, "loss": 6.71, "step": 30 }, { "epoch": 0.015421351109342354, "grad_norm": 1.2283194065093994, "learning_rate": 0.00019845405110904146, "loss": 6.6504, "step": 31 }, { "epoch": 0.015918814048353398, "grad_norm": 0.8201404213905334, "learning_rate": 0.00019246656176210558, "loss": 6.7202, "step": 32 }, { "epoch": 0.01641627698736444, "grad_norm": 0.8443421125411987, "learning_rate": 0.0001864221437715939, "loss": 6.8298, "step": 33 }, { "epoch": 0.016913739926375484, "grad_norm": 1.0041660070419312, "learning_rate": 0.0001803333251168141, "loss": 6.8294, "step": 34 }, { "epoch": 0.01741120286538653, "grad_norm": 0.729718804359436, "learning_rate": 0.00017421272580423058, "loss": 6.7173, "step": 35 }, { "epoch": 0.017908665804397574, "grad_norm": 0.9453200101852417, "learning_rate": 0.00016807303171057425, "loss": 6.6455, "step": 36 }, { "epoch": 0.018406128743408617, "grad_norm": 1.152786135673523, "learning_rate": 0.00016192696828942573, "loss": 6.5751, "step": 37 }, { "epoch": 0.01890359168241966, "grad_norm": 1.0720747709274292, "learning_rate": 0.00015578727419576942, "loss": 6.8176, "step": 38 }, { "epoch": 0.019401054621430703, "grad_norm": 1.0019649267196655, "learning_rate": 0.00014966667488318586, "loss": 6.8212, "step": 39 }, { "epoch": 0.019898517560441746, "grad_norm": 1.0470595359802246, "learning_rate": 0.00014357785622840606, "loss": 6.8254, "step": 40 }, { "epoch": 0.02039598049945279, "grad_norm": 1.326593279838562, "learning_rate": 0.00013753343823789445, "loss": 6.7906, "step": 41 }, { "epoch": 0.020893443438463836, "grad_norm": 0.5931538939476013, "learning_rate": 0.00013154594889095854, "loss": 6.7336, "step": 42 }, { "epoch": 0.02139090637747488, "grad_norm": 0.9888867735862732, "learning_rate": 0.00012562779817359865, "loss": 6.6365, "step": 43 }, { "epoch": 0.021888369316485922, "grad_norm": 1.1606509685516357, "learning_rate": 0.00011979125235691685, "loss": 6.5585, "step": 44 }, { "epoch": 0.022385832255496965, "grad_norm": 0.9471900463104248, "learning_rate": 0.00011404840857339706, "loss": 6.7794, "step": 45 }, { "epoch": 0.02288329519450801, "grad_norm": 1.0708708763122559, "learning_rate": 0.0001084111697437504, "loss": 6.8215, "step": 46 }, { "epoch": 0.02338075813351905, "grad_norm": 0.7427545785903931, "learning_rate": 0.00010289121990629447, "loss": 6.7288, "step": 47 }, { "epoch": 0.023878221072530098, "grad_norm": 0.9723594188690186, "learning_rate": 9.750000000000003e-05, "loss": 6.6319, "step": 48 }, { "epoch": 0.02437568401154114, "grad_norm": 0.9104545712471008, "learning_rate": 9.22486841513973e-05, "loss": 6.5588, "step": 49 }, { "epoch": 0.024873146950552184, "grad_norm": 0.8541779518127441, "learning_rate": 8.714815651449293e-05, "loss": 6.6375, "step": 50 }, { "epoch": 0.024873146950552184, "eval_loss": 6.703342437744141, "eval_runtime": 0.0414, "eval_samples_per_second": 1208.754, "eval_steps_per_second": 72.525, "step": 50 }, { "epoch": 0.025370609889563227, "grad_norm": 1.8086193799972534, "learning_rate": 8.220898871169827e-05, "loss": 6.8079, "step": 51 }, { "epoch": 0.02586807282857427, "grad_norm": 0.7370948195457458, "learning_rate": 7.744141792252794e-05, "loss": 6.8023, "step": 52 }, { "epoch": 0.026365535767585314, "grad_norm": 1.0690277814865112, "learning_rate": 7.285532566548172e-05, "loss": 6.7905, "step": 53 }, { "epoch": 0.026862998706596357, "grad_norm": 1.2137597799301147, "learning_rate": 6.846021731708856e-05, "loss": 6.77, "step": 54 }, { "epoch": 0.027360461645607403, "grad_norm": 0.7531571984291077, "learning_rate": 6.426520241056245e-05, "loss": 6.661, "step": 55 }, { "epoch": 0.027857924584618447, "grad_norm": 0.8767682313919067, "learning_rate": 6.0278975754903317e-05, "loss": 6.6017, "step": 56 }, { "epoch": 0.02835538752362949, "grad_norm": 0.7823953628540039, "learning_rate": 5.6509799413577934e-05, "loss": 6.5955, "step": 57 }, { "epoch": 0.028852850462640533, "grad_norm": 0.8727318048477173, "learning_rate": 5.296548558013161e-05, "loss": 6.7922, "step": 58 }, { "epoch": 0.029350313401651576, "grad_norm": 0.996519148349762, "learning_rate": 4.9653380386224046e-05, "loss": 6.7956, "step": 59 }, { "epoch": 0.02984777634066262, "grad_norm": 0.6733922362327576, "learning_rate": 4.658034867564977e-05, "loss": 6.6694, "step": 60 }, { "epoch": 0.030345239279673666, "grad_norm": 0.8655278086662292, "learning_rate": 4.375275977590104e-05, "loss": 6.5892, "step": 61 }, { "epoch": 0.03084270221868471, "grad_norm": 0.9837325811386108, "learning_rate": 4.117647429676387e-05, "loss": 6.5165, "step": 62 }, { "epoch": 0.03134016515769575, "grad_norm": 1.1237236261367798, "learning_rate": 3.885683198330941e-05, "loss": 6.7888, "step": 63 }, { "epoch": 0.031837628096706795, "grad_norm": 1.0909230709075928, "learning_rate": 3.679864064845691e-05, "loss": 6.8041, "step": 64 }, { "epoch": 0.03233509103571784, "grad_norm": 0.9223988652229309, "learning_rate": 3.500616620804712e-05, "loss": 6.7929, "step": 65 }, { "epoch": 0.03283255397472888, "grad_norm": 1.2748134136199951, "learning_rate": 3.348312383908033e-05, "loss": 6.7633, "step": 66 }, { "epoch": 0.03333001691373993, "grad_norm": 0.5331524014472961, "learning_rate": 3.223267027944483e-05, "loss": 6.7201, "step": 67 }, { "epoch": 0.03382747985275097, "grad_norm": 0.9274587035179138, "learning_rate": 3.125739728509535e-05, "loss": 6.6011, "step": 68 }, { "epoch": 0.034324942791762014, "grad_norm": 1.011277198791504, "learning_rate": 3.055932625824328e-05, "loss": 6.5109, "step": 69 }, { "epoch": 0.03482240573077306, "grad_norm": 0.7799229025840759, "learning_rate": 3.0139904057691777e-05, "loss": 6.6674, "step": 70 }, { "epoch": 0.0353198686697841, "grad_norm": 1.039928674697876, "learning_rate": 2.9999999999999997e-05, "loss": 6.8205, "step": 71 } ], "logging_steps": 1, "max_steps": 71, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 1, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7767234478080.0, "train_batch_size": 6, "trial_name": null, "trial_params": null }