{ "best_metric": 2.4195876121520996, "best_model_checkpoint": "./output/training_results/C016_random_sample_llama3-8b-base_pretrain_20240504_181744/checkpoint-1000", "epoch": 4.0, "eval_steps": 200, "global_step": 4108, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009737098344693282, "grad_norm": 4.622597308389442, "learning_rate": 7.5e-07, "loss": 2.7686, "step": 1 }, { "epoch": 0.004868549172346641, "grad_norm": 2.9010810501364586, "learning_rate": 3e-06, "loss": 2.7274, "step": 5 }, { "epoch": 0.009737098344693282, "grad_norm": 2.6032598047562, "learning_rate": 6e-06, "loss": 2.667, "step": 10 }, { "epoch": 0.014605647517039922, "grad_norm": 2.570724795196223, "learning_rate": 9.75e-06, "loss": 2.6551, "step": 15 }, { "epoch": 0.019474196689386564, "grad_norm": 2.968267308985469, "learning_rate": 1.3500000000000001e-05, "loss": 2.643, "step": 20 }, { "epoch": 0.024342745861733205, "grad_norm": 2.8739885340338285, "learning_rate": 1.487975935855361e-05, "loss": 2.6294, "step": 25 }, { "epoch": 0.029211295034079845, "grad_norm": 3.023512383227986, "learning_rate": 1.4681310547763171e-05, "loss": 2.627, "step": 30 }, { "epoch": 0.034079844206426485, "grad_norm": 2.528499970351483, "learning_rate": 1.4485278864193794e-05, "loss": 2.5955, "step": 35 }, { "epoch": 0.03894839337877313, "grad_norm": 2.547095976749082, "learning_rate": 1.4291637778717775e-05, "loss": 2.6183, "step": 40 }, { "epoch": 0.043816942551119765, "grad_norm": 1.9843941555908906, "learning_rate": 1.4100361021341427e-05, "loss": 2.5925, "step": 45 }, { "epoch": 0.04868549172346641, "grad_norm": 2.0711055751945815, "learning_rate": 1.3911422578987613e-05, "loss": 2.5762, "step": 50 }, { "epoch": 0.053554040895813046, "grad_norm": 2.1149009622981736, "learning_rate": 1.3724796693294575e-05, "loss": 2.6115, "step": 55 }, { "epoch": 0.05842259006815969, "grad_norm": 2.125830374210055, "learning_rate": 1.3540457858430796e-05, "loss": 2.6139, "step": 60 }, { "epoch": 0.06329113924050633, "grad_norm": 1.9410826947795146, "learning_rate": 1.3394616487103794e-05, "loss": 2.5748, "step": 65 }, { "epoch": 0.06815968841285297, "grad_norm": 2.1512046862388146, "learning_rate": 1.3214330866815837e-05, "loss": 2.6165, "step": 70 }, { "epoch": 0.0730282375851996, "grad_norm": 2.0456636080847157, "learning_rate": 1.3036262206089142e-05, "loss": 2.529, "step": 75 }, { "epoch": 0.07789678675754626, "grad_norm": 2.232330041514811, "learning_rate": 1.2860385939062166e-05, "loss": 2.6501, "step": 80 }, { "epoch": 0.0827653359298929, "grad_norm": 2.415240950467998, "learning_rate": 1.2686677742137694e-05, "loss": 2.5821, "step": 85 }, { "epoch": 0.08763388510223953, "grad_norm": 2.4124279122793375, "learning_rate": 1.2515113531889656e-05, "loss": 2.5964, "step": 90 }, { "epoch": 0.09250243427458617, "grad_norm": 2.2985857393283364, "learning_rate": 1.2345669462985584e-05, "loss": 2.6227, "step": 95 }, { "epoch": 0.09737098344693282, "grad_norm": 1.984056163436905, "learning_rate": 1.2178321926124443e-05, "loss": 2.5646, "step": 100 }, { "epoch": 0.10223953261927946, "grad_norm": 1.9469124194664196, "learning_rate": 1.201304754598983e-05, "loss": 2.5677, "step": 105 }, { "epoch": 0.10710808179162609, "grad_norm": 2.000422025073212, "learning_rate": 1.1849823179218302e-05, "loss": 2.6047, "step": 110 }, { "epoch": 0.11197663096397274, "grad_norm": 1.8457596949821262, "learning_rate": 1.168862591238295e-05, "loss": 2.6201, "step": 115 }, { "epoch": 0.11684518013631938, "grad_norm": 1.8876145847065753, "learning_rate": 1.1529433059992029e-05, "loss": 2.5327, "step": 120 }, { "epoch": 0.12171372930866602, "grad_norm": 1.8657408134698463, "learning_rate": 1.1372222162502398e-05, "loss": 2.5759, "step": 125 }, { "epoch": 0.12658227848101267, "grad_norm": 1.9689520876834024, "learning_rate": 1.1216970984347951e-05, "loss": 2.5822, "step": 130 }, { "epoch": 0.1314508276533593, "grad_norm": 1.8787257890744393, "learning_rate": 1.1063657511982824e-05, "loss": 2.5703, "step": 135 }, { "epoch": 0.13631937682570594, "grad_norm": 2.3946720962934673, "learning_rate": 1.0912259951939131e-05, "loss": 2.5048, "step": 140 }, { "epoch": 0.14118792599805258, "grad_norm": 2.069484826525082, "learning_rate": 1.0762756728899413e-05, "loss": 2.5433, "step": 145 }, { "epoch": 0.1460564751703992, "grad_norm": 1.9728419903336738, "learning_rate": 1.0615126483783578e-05, "loss": 2.5451, "step": 150 }, { "epoch": 0.15092502434274585, "grad_norm": 1.8965209810159926, "learning_rate": 1.0469348071850116e-05, "loss": 2.5761, "step": 155 }, { "epoch": 0.15579357351509251, "grad_norm": 1.9728913609239898, "learning_rate": 1.0325400560811745e-05, "loss": 2.5509, "step": 160 }, { "epoch": 0.16066212268743915, "grad_norm": 1.7695814333016529, "learning_rate": 1.0183263228965269e-05, "loss": 2.5695, "step": 165 }, { "epoch": 0.1655306718597858, "grad_norm": 1.9109787480384295, "learning_rate": 1.0042915563335448e-05, "loss": 2.5843, "step": 170 }, { "epoch": 0.17039922103213243, "grad_norm": 1.8950613853479759, "learning_rate": 9.904337257833037e-06, "loss": 2.5447, "step": 175 }, { "epoch": 0.17526777020447906, "grad_norm": 1.9462634859013892, "learning_rate": 9.767508211426758e-06, "loss": 2.5236, "step": 180 }, { "epoch": 0.1801363193768257, "grad_norm": 1.9007929646893995, "learning_rate": 9.632408526329031e-06, "loss": 2.5511, "step": 185 }, { "epoch": 0.18500486854917234, "grad_norm": 1.9203529678658229, "learning_rate": 9.4990185061956e-06, "loss": 2.5294, "step": 190 }, { "epoch": 0.189873417721519, "grad_norm": 1.8492777850666573, "learning_rate": 9.367318654338783e-06, "loss": 2.5313, "step": 195 }, { "epoch": 0.19474196689386564, "grad_norm": 2.053277986812429, "learning_rate": 9.23728967195438e-06, "loss": 2.5472, "step": 200 }, { "epoch": 0.19474196689386564, "eval_loss": 2.526153564453125, "eval_runtime": 86.8653, "eval_samples_per_second": 84.027, "eval_steps_per_second": 0.668, "step": 200 }, { "epoch": 0.19961051606621227, "grad_norm": 1.8721064281522946, "learning_rate": 9.108912456361987e-06, "loss": 2.5221, "step": 205 }, { "epoch": 0.2044790652385589, "grad_norm": 2.0199844942072707, "learning_rate": 8.982168099258849e-06, "loss": 2.5093, "step": 210 }, { "epoch": 0.20934761441090555, "grad_norm": 1.7890996151627292, "learning_rate": 8.857037884987036e-06, "loss": 2.5197, "step": 215 }, { "epoch": 0.21421616358325218, "grad_norm": 1.8519229946590852, "learning_rate": 8.733503288813807e-06, "loss": 2.4945, "step": 220 }, { "epoch": 0.21908471275559882, "grad_norm": 1.7635414105359053, "learning_rate": 8.611545975225206e-06, "loss": 2.508, "step": 225 }, { "epoch": 0.22395326192794549, "grad_norm": 1.8628950690860595, "learning_rate": 8.491147796232788e-06, "loss": 2.5093, "step": 230 }, { "epoch": 0.22882181110029212, "grad_norm": 1.8308684769690282, "learning_rate": 8.372290789693232e-06, "loss": 2.4912, "step": 235 }, { "epoch": 0.23369036027263876, "grad_norm": 1.8771495332116028, "learning_rate": 8.254957177640966e-06, "loss": 2.5112, "step": 240 }, { "epoch": 0.2385589094449854, "grad_norm": 1.848114881393312, "learning_rate": 8.13912936463366e-06, "loss": 2.5004, "step": 245 }, { "epoch": 0.24342745861733203, "grad_norm": 1.8532877282661944, "learning_rate": 8.024789936110332e-06, "loss": 2.5063, "step": 250 }, { "epoch": 0.24829600778967867, "grad_norm": 1.8072290437694474, "learning_rate": 7.911921656762299e-06, "loss": 2.4812, "step": 255 }, { "epoch": 0.25316455696202533, "grad_norm": 1.7086014592127396, "learning_rate": 7.800507468916596e-06, "loss": 2.5107, "step": 260 }, { "epoch": 0.25803310613437197, "grad_norm": 1.713632600119656, "learning_rate": 7.690530490932004e-06, "loss": 2.4908, "step": 265 }, { "epoch": 0.2629016553067186, "grad_norm": 1.7625004775631927, "learning_rate": 7.581974015607404e-06, "loss": 2.5364, "step": 270 }, { "epoch": 0.26777020447906524, "grad_norm": 1.73225200198116, "learning_rate": 7.474821508602552e-06, "loss": 2.4639, "step": 275 }, { "epoch": 0.2726387536514119, "grad_norm": 1.7849989992249207, "learning_rate": 7.369056606871157e-06, "loss": 2.5102, "step": 280 }, { "epoch": 0.2775073028237585, "grad_norm": 1.8370989545866505, "learning_rate": 7.2646631171060325e-06, "loss": 2.4668, "step": 285 }, { "epoch": 0.28237585199610515, "grad_norm": 1.7638888058771318, "learning_rate": 7.161625014196459e-06, "loss": 2.4778, "step": 290 }, { "epoch": 0.2872444011684518, "grad_norm": 1.840487665783078, "learning_rate": 7.059926439697582e-06, "loss": 2.4949, "step": 295 }, { "epoch": 0.2921129503407984, "grad_norm": 1.7654501649580268, "learning_rate": 6.959551700311679e-06, "loss": 2.5254, "step": 300 }, { "epoch": 0.29698149951314506, "grad_norm": 1.9353359489811188, "learning_rate": 6.860485266381383e-06, "loss": 2.4986, "step": 305 }, { "epoch": 0.3018500486854917, "grad_norm": 1.9617969898380334, "learning_rate": 6.7627117703947036e-06, "loss": 2.5186, "step": 310 }, { "epoch": 0.30671859785783834, "grad_norm": 1.702196140955499, "learning_rate": 6.6662160055017e-06, "loss": 2.4988, "step": 315 }, { "epoch": 0.31158714703018503, "grad_norm": 1.6844416023733326, "learning_rate": 6.570982924042871e-06, "loss": 2.5118, "step": 320 }, { "epoch": 0.31645569620253167, "grad_norm": 1.7294022438189514, "learning_rate": 6.476997636089112e-06, "loss": 2.5248, "step": 325 }, { "epoch": 0.3213242453748783, "grad_norm": 2.0648141831567983, "learning_rate": 6.384245407993103e-06, "loss": 2.5276, "step": 330 }, { "epoch": 0.32619279454722494, "grad_norm": 1.728780446047325, "learning_rate": 6.292711660952165e-06, "loss": 2.4661, "step": 335 }, { "epoch": 0.3310613437195716, "grad_norm": 1.7061810945106262, "learning_rate": 6.202381969582487e-06, "loss": 2.4577, "step": 340 }, { "epoch": 0.3359298928919182, "grad_norm": 1.6870860598605661, "learning_rate": 6.11324206050453e-06, "loss": 2.4968, "step": 345 }, { "epoch": 0.34079844206426485, "grad_norm": 1.7294693647713457, "learning_rate": 6.025277810939709e-06, "loss": 2.4665, "step": 350 }, { "epoch": 0.3456669912366115, "grad_norm": 1.771451497626332, "learning_rate": 5.9384752473181495e-06, "loss": 2.4601, "step": 355 }, { "epoch": 0.3505355404089581, "grad_norm": 1.7336392065891828, "learning_rate": 5.8528205438975404e-06, "loss": 2.4733, "step": 360 }, { "epoch": 0.35540408958130476, "grad_norm": 1.751587409627844, "learning_rate": 5.768300021392871e-06, "loss": 2.4184, "step": 365 }, { "epoch": 0.3602726387536514, "grad_norm": 1.735366924052735, "learning_rate": 5.684900145617154e-06, "loss": 2.4582, "step": 370 }, { "epoch": 0.36514118792599803, "grad_norm": 1.8313267360344203, "learning_rate": 5.6026075261329595e-06, "loss": 2.4746, "step": 375 }, { "epoch": 0.37000973709834467, "grad_norm": 1.8199109237326785, "learning_rate": 5.521408914914653e-06, "loss": 2.4395, "step": 380 }, { "epoch": 0.3748782862706913, "grad_norm": 1.655319976322653, "learning_rate": 5.441291205021376e-06, "loss": 2.4805, "step": 385 }, { "epoch": 0.379746835443038, "grad_norm": 1.6994219665911234, "learning_rate": 5.3622414292806385e-06, "loss": 2.4505, "step": 390 }, { "epoch": 0.38461538461538464, "grad_norm": 1.7131089370429131, "learning_rate": 5.2842467589823945e-06, "loss": 2.5166, "step": 395 }, { "epoch": 0.3894839337877313, "grad_norm": 1.7028942553739241, "learning_rate": 5.207294502583657e-06, "loss": 2.4431, "step": 400 }, { "epoch": 0.3894839337877313, "eval_loss": 2.4732775688171387, "eval_runtime": 85.4562, "eval_samples_per_second": 85.412, "eval_steps_per_second": 0.679, "step": 400 }, { "epoch": 0.3943524829600779, "grad_norm": 1.772274911090255, "learning_rate": 5.131372104423501e-06, "loss": 2.5029, "step": 405 }, { "epoch": 0.39922103213242455, "grad_norm": 1.6531175372154208, "learning_rate": 5.0564671434483495e-06, "loss": 2.5081, "step": 410 }, { "epoch": 0.4040895813047712, "grad_norm": 1.649280502586864, "learning_rate": 4.9825673319475865e-06, "loss": 2.4378, "step": 415 }, { "epoch": 0.4089581304771178, "grad_norm": 1.8019468764704318, "learning_rate": 4.909660514299323e-06, "loss": 2.5013, "step": 420 }, { "epoch": 0.41382667964946446, "grad_norm": 2.001260057759529, "learning_rate": 4.837734665726331e-06, "loss": 2.4586, "step": 425 }, { "epoch": 0.4186952288218111, "grad_norm": 1.7081822720906132, "learning_rate": 4.766777891061954e-06, "loss": 2.513, "step": 430 }, { "epoch": 0.42356377799415773, "grad_norm": 1.8238918791581111, "learning_rate": 4.696778423526121e-06, "loss": 2.4336, "step": 435 }, { "epoch": 0.42843232716650437, "grad_norm": 1.7162241591421625, "learning_rate": 4.627724623511167e-06, "loss": 2.4923, "step": 440 }, { "epoch": 0.433300876338851, "grad_norm": 1.791273570594169, "learning_rate": 4.559604977377591e-06, "loss": 2.4524, "step": 445 }, { "epoch": 0.43816942551119764, "grad_norm": 1.585355030794431, "learning_rate": 4.4924080962595615e-06, "loss": 2.4547, "step": 450 }, { "epoch": 0.4430379746835443, "grad_norm": 1.627715490257672, "learning_rate": 4.426122714880177e-06, "loss": 2.4271, "step": 455 }, { "epoch": 0.44790652385589097, "grad_norm": 1.7001533187863673, "learning_rate": 4.360737690376327e-06, "loss": 2.5004, "step": 460 }, { "epoch": 0.4527750730282376, "grad_norm": 1.645312039241492, "learning_rate": 4.29624200113319e-06, "loss": 2.4227, "step": 465 }, { "epoch": 0.45764362220058424, "grad_norm": 1.6935805821575964, "learning_rate": 4.232624745628264e-06, "loss": 2.4804, "step": 470 }, { "epoch": 0.4625121713729309, "grad_norm": 1.6013179989515345, "learning_rate": 4.169875141284801e-06, "loss": 2.4438, "step": 475 }, { "epoch": 0.4673807205452775, "grad_norm": 1.6369493229279692, "learning_rate": 4.107982523334686e-06, "loss": 2.468, "step": 480 }, { "epoch": 0.47224926971762415, "grad_norm": 1.5838983294803852, "learning_rate": 4.046936343690647e-06, "loss": 2.4468, "step": 485 }, { "epoch": 0.4771178188899708, "grad_norm": 1.652021178997152, "learning_rate": 3.986726169827688e-06, "loss": 2.4484, "step": 490 }, { "epoch": 0.4819863680623174, "grad_norm": 1.6100438156106431, "learning_rate": 3.9391530173378065e-06, "loss": 2.4699, "step": 495 }, { "epoch": 0.48685491723466406, "grad_norm": 1.6357278449106312, "learning_rate": 3.880421728549051e-06, "loss": 2.4972, "step": 500 }, { "epoch": 0.4917234664070107, "grad_norm": 1.665827366459226, "learning_rate": 3.822497839855098e-06, "loss": 2.4498, "step": 505 }, { "epoch": 0.49659201557935734, "grad_norm": 1.6865163895540334, "learning_rate": 3.7653713478280894e-06, "loss": 2.4344, "step": 510 }, { "epoch": 0.501460564751704, "grad_norm": 1.6041809903831323, "learning_rate": 3.70903235936047e-06, "loss": 2.4389, "step": 515 }, { "epoch": 0.5063291139240507, "grad_norm": 1.5953703380549231, "learning_rate": 3.653471090598976e-06, "loss": 2.456, "step": 520 }, { "epoch": 0.5111976630963972, "grad_norm": 1.651511590173226, "learning_rate": 3.59867786588742e-06, "loss": 2.4241, "step": 525 }, { "epoch": 0.5160662122687439, "grad_norm": 1.7103129608293084, "learning_rate": 3.5446431167182903e-06, "loss": 2.4655, "step": 530 }, { "epoch": 0.5209347614410905, "grad_norm": 1.6921307716416762, "learning_rate": 3.49135738069308e-06, "loss": 2.4682, "step": 535 }, { "epoch": 0.5258033106134372, "grad_norm": 2.4654732646603064, "learning_rate": 3.438811300491226e-06, "loss": 2.4509, "step": 540 }, { "epoch": 0.5306718597857838, "grad_norm": 1.8390744205983383, "learning_rate": 3.386995622847693e-06, "loss": 2.4649, "step": 545 }, { "epoch": 0.5355404089581305, "grad_norm": 1.644182617363712, "learning_rate": 3.3359011975390947e-06, "loss": 2.4421, "step": 550 }, { "epoch": 0.5404089581304771, "grad_norm": 1.6574734258059267, "learning_rate": 3.2855189763782522e-06, "loss": 2.4339, "step": 555 }, { "epoch": 0.5452775073028238, "grad_norm": 1.645146888773675, "learning_rate": 3.2358400122172194e-06, "loss": 2.4573, "step": 560 }, { "epoch": 0.5501460564751705, "grad_norm": 1.6192183546217223, "learning_rate": 3.1868554579586605e-06, "loss": 2.4574, "step": 565 }, { "epoch": 0.555014605647517, "grad_norm": 1.671822469835026, "learning_rate": 3.1385565655755006e-06, "loss": 2.4685, "step": 570 }, { "epoch": 0.5598831548198637, "grad_norm": 1.6395167213969633, "learning_rate": 3.0909346851388586e-06, "loss": 2.4845, "step": 575 }, { "epoch": 0.5647517039922103, "grad_norm": 1.6945968363900514, "learning_rate": 3.0439812638541433e-06, "loss": 2.4732, "step": 580 }, { "epoch": 0.569620253164557, "grad_norm": 1.6428936915822159, "learning_rate": 2.9976878451053044e-06, "loss": 2.4622, "step": 585 }, { "epoch": 0.5744888023369036, "grad_norm": 1.6443646175732205, "learning_rate": 2.9520460675071095e-06, "loss": 2.4466, "step": 590 }, { "epoch": 0.5793573515092503, "grad_norm": 1.6462331908034773, "learning_rate": 2.907047663965498e-06, "loss": 2.466, "step": 595 }, { "epoch": 0.5842259006815969, "grad_norm": 1.6579152402369843, "learning_rate": 2.862684460745835e-06, "loss": 2.4163, "step": 600 }, { "epoch": 0.5842259006815969, "eval_loss": 2.4442696571350098, "eval_runtime": 85.0459, "eval_samples_per_second": 85.824, "eval_steps_per_second": 0.682, "step": 600 }, { "epoch": 0.5890944498539435, "grad_norm": 1.6546825117019377, "learning_rate": 2.8189483765491077e-06, "loss": 2.4857, "step": 605 }, { "epoch": 0.5939629990262901, "grad_norm": 1.6868467390816666, "learning_rate": 2.775831421595948e-06, "loss": 2.4588, "step": 610 }, { "epoch": 0.5988315481986368, "grad_norm": 1.631864642870465, "learning_rate": 2.7333256967184713e-06, "loss": 2.4542, "step": 615 }, { "epoch": 0.6037000973709834, "grad_norm": 1.5765796237275709, "learning_rate": 2.6914233924598045e-06, "loss": 2.4277, "step": 620 }, { "epoch": 0.6085686465433301, "grad_norm": 1.6329988363272994, "learning_rate": 2.6501167881813393e-06, "loss": 2.3847, "step": 625 }, { "epoch": 0.6134371957156767, "grad_norm": 1.6799124493504514, "learning_rate": 2.6093982511776034e-06, "loss": 2.4285, "step": 630 }, { "epoch": 0.6183057448880234, "grad_norm": 1.6239504885103953, "learning_rate": 2.569260235798673e-06, "loss": 2.4432, "step": 635 }, { "epoch": 0.6231742940603701, "grad_norm": 1.6009220872039354, "learning_rate": 2.5296952825801428e-06, "loss": 2.4118, "step": 640 }, { "epoch": 0.6280428432327166, "grad_norm": 1.668850128404625, "learning_rate": 2.49069601738053e-06, "loss": 2.4968, "step": 645 }, { "epoch": 0.6329113924050633, "grad_norm": 1.6629353311695643, "learning_rate": 2.45225515052612e-06, "loss": 2.4499, "step": 650 }, { "epoch": 0.6377799415774099, "grad_norm": 1.5929095731919567, "learning_rate": 2.414365475963123e-06, "loss": 2.4344, "step": 655 }, { "epoch": 0.6426484907497566, "grad_norm": 1.6171386676286523, "learning_rate": 2.3770198704171897e-06, "loss": 2.4291, "step": 660 }, { "epoch": 0.6475170399221032, "grad_norm": 1.6420238027175638, "learning_rate": 2.3402112925601284e-06, "loss": 2.4089, "step": 665 }, { "epoch": 0.6523855890944499, "grad_norm": 1.6788147362161117, "learning_rate": 2.30393278218386e-06, "loss": 2.4573, "step": 670 }, { "epoch": 0.6572541382667965, "grad_norm": 1.6870203522696865, "learning_rate": 2.2681774593814984e-06, "loss": 2.4088, "step": 675 }, { "epoch": 0.6621226874391432, "grad_norm": 1.635603955024262, "learning_rate": 2.232938523735545e-06, "loss": 2.4588, "step": 680 }, { "epoch": 0.6669912366114897, "grad_norm": 1.6513072405624771, "learning_rate": 2.1982092535131004e-06, "loss": 2.4498, "step": 685 }, { "epoch": 0.6718597857838364, "grad_norm": 1.6924700394115, "learning_rate": 2.1639830048681073e-06, "loss": 2.4365, "step": 690 }, { "epoch": 0.676728334956183, "grad_norm": 1.6617480593936398, "learning_rate": 2.130253211050499e-06, "loss": 2.4644, "step": 695 }, { "epoch": 0.6815968841285297, "grad_norm": 1.6227857828645276, "learning_rate": 2.097013381622268e-06, "loss": 2.4065, "step": 700 }, { "epoch": 0.6864654333008764, "grad_norm": 1.6566150548613952, "learning_rate": 2.0642571016803623e-06, "loss": 2.4368, "step": 705 }, { "epoch": 0.691333982473223, "grad_norm": 2.0603714204760255, "learning_rate": 2.031978031086392e-06, "loss": 2.4595, "step": 710 }, { "epoch": 0.6962025316455697, "grad_norm": 1.6243505097480613, "learning_rate": 2.0001699037030468e-06, "loss": 2.4347, "step": 715 }, { "epoch": 0.7010710808179162, "grad_norm": 1.7249749503578298, "learning_rate": 1.968826526637239e-06, "loss": 2.4648, "step": 720 }, { "epoch": 0.7059396299902629, "grad_norm": 1.7262462882996954, "learning_rate": 1.9379417794898855e-06, "loss": 2.439, "step": 725 }, { "epoch": 0.7108081791626095, "grad_norm": 1.6204017543437201, "learning_rate": 1.907509613612265e-06, "loss": 2.4204, "step": 730 }, { "epoch": 0.7156767283349562, "grad_norm": 1.6204489322103508, "learning_rate": 1.8775240513689499e-06, "loss": 2.4594, "step": 735 }, { "epoch": 0.7205452775073028, "grad_norm": 1.6106742671779046, "learning_rate": 1.8479791854072195e-06, "loss": 2.4215, "step": 740 }, { "epoch": 0.7254138266796495, "grad_norm": 1.6278702344437153, "learning_rate": 1.818869177932943e-06, "loss": 2.4177, "step": 745 }, { "epoch": 0.7302823758519961, "grad_norm": 1.702303345942966, "learning_rate": 1.7901882599928386e-06, "loss": 2.4412, "step": 750 }, { "epoch": 0.7351509250243428, "grad_norm": 1.616327757952299, "learning_rate": 1.7619307307631266e-06, "loss": 2.4, "step": 755 }, { "epoch": 0.7400194741966893, "grad_norm": 1.581113737981139, "learning_rate": 1.734090956844456e-06, "loss": 2.4072, "step": 760 }, { "epoch": 0.744888023369036, "grad_norm": 1.556080290425103, "learning_rate": 1.7066633715631195e-06, "loss": 2.4384, "step": 765 }, { "epoch": 0.7497565725413826, "grad_norm": 1.6038706235589455, "learning_rate": 1.6796424742784764e-06, "loss": 2.4738, "step": 770 }, { "epoch": 0.7546251217137293, "grad_norm": 1.662078041711206, "learning_rate": 1.6530228296965529e-06, "loss": 2.4957, "step": 775 }, { "epoch": 0.759493670886076, "grad_norm": 1.6328340091897662, "learning_rate": 1.6267990671897495e-06, "loss": 2.4816, "step": 780 }, { "epoch": 0.7643622200584226, "grad_norm": 1.6438111553611983, "learning_rate": 1.6009658801226467e-06, "loss": 2.4171, "step": 785 }, { "epoch": 0.7692307692307693, "grad_norm": 1.7211490194262304, "learning_rate": 1.575518025183845e-06, "loss": 2.4268, "step": 790 }, { "epoch": 0.7740993184031159, "grad_norm": 1.7200203924352186, "learning_rate": 1.5504503217237718e-06, "loss": 2.4661, "step": 795 }, { "epoch": 0.7789678675754625, "grad_norm": 1.6451751546473583, "learning_rate": 1.5257576510984581e-06, "loss": 2.4462, "step": 800 }, { "epoch": 0.7789678675754625, "eval_loss": 2.4281296730041504, "eval_runtime": 85.85, "eval_samples_per_second": 85.02, "eval_steps_per_second": 0.676, "step": 800 }, { "epoch": 0.7838364167478091, "grad_norm": 1.7186162803254572, "learning_rate": 1.5014349560191975e-06, "loss": 2.4447, "step": 805 }, { "epoch": 0.7887049659201558, "grad_norm": 1.6390532449357682, "learning_rate": 1.4774772399080773e-06, "loss": 2.3768, "step": 810 }, { "epoch": 0.7935735150925024, "grad_norm": 1.6479758648696174, "learning_rate": 1.4538795662592943e-06, "loss": 2.4559, "step": 815 }, { "epoch": 0.7984420642648491, "grad_norm": 1.6376458723290501, "learning_rate": 1.4306370580062715e-06, "loss": 2.4113, "step": 820 }, { "epoch": 0.8033106134371957, "grad_norm": 1.6642605682645168, "learning_rate": 1.4077448968944671e-06, "loss": 2.4438, "step": 825 }, { "epoch": 0.8081791626095424, "grad_norm": 1.5597665460776493, "learning_rate": 1.3851983228598879e-06, "loss": 2.4615, "step": 830 }, { "epoch": 0.813047711781889, "grad_norm": 1.5998155659125195, "learning_rate": 1.3629926334132263e-06, "loss": 2.4245, "step": 835 }, { "epoch": 0.8179162609542356, "grad_norm": 1.6231320102269173, "learning_rate": 1.341123183029607e-06, "loss": 2.4677, "step": 840 }, { "epoch": 0.8227848101265823, "grad_norm": 1.587206597114467, "learning_rate": 1.3195853825438637e-06, "loss": 2.4268, "step": 845 }, { "epoch": 0.8276533592989289, "grad_norm": 1.69258641098763, "learning_rate": 1.2983746985513552e-06, "loss": 2.4468, "step": 850 }, { "epoch": 0.8325219084712756, "grad_norm": 1.6329982694488412, "learning_rate": 1.2774866528142222e-06, "loss": 2.4483, "step": 855 }, { "epoch": 0.8373904576436222, "grad_norm": 1.6510229207341078, "learning_rate": 1.2569168216730917e-06, "loss": 2.4192, "step": 860 }, { "epoch": 0.8422590068159689, "grad_norm": 1.5817582463558826, "learning_rate": 1.2366608354641566e-06, "loss": 2.4214, "step": 865 }, { "epoch": 0.8471275559883155, "grad_norm": 2.4364871892272117, "learning_rate": 1.216714377941609e-06, "loss": 2.3713, "step": 870 }, { "epoch": 0.8519961051606622, "grad_norm": 2.5372621502402186, "learning_rate": 1.197073185705361e-06, "loss": 2.415, "step": 875 }, { "epoch": 0.8568646543330087, "grad_norm": 1.7440355879318754, "learning_rate": 1.1777330476340439e-06, "loss": 2.4468, "step": 880 }, { "epoch": 0.8617332035053554, "grad_norm": 1.5959852776726504, "learning_rate": 1.1624748993325157e-06, "loss": 2.4416, "step": 885 }, { "epoch": 0.866601752677702, "grad_norm": 1.5792041005336184, "learning_rate": 1.1436662114572268e-06, "loss": 2.4699, "step": 890 }, { "epoch": 0.8714703018500487, "grad_norm": 1.6749035075354037, "learning_rate": 1.1251470599677525e-06, "loss": 2.3906, "step": 895 }, { "epoch": 0.8763388510223953, "grad_norm": 1.6974831202238516, "learning_rate": 1.106913427322351e-06, "loss": 2.4473, "step": 900 }, { "epoch": 0.881207400194742, "grad_norm": 1.629005550261729, "learning_rate": 1.0889613456083674e-06, "loss": 2.3865, "step": 905 }, { "epoch": 0.8860759493670886, "grad_norm": 1.541318076309618, "learning_rate": 1.0712868960049544e-06, "loss": 2.4382, "step": 910 }, { "epoch": 0.8909444985394352, "grad_norm": 1.5752275103135045, "learning_rate": 1.0538862082508074e-06, "loss": 2.452, "step": 915 }, { "epoch": 0.8958130477117819, "grad_norm": 1.699492645238764, "learning_rate": 1.036755460116832e-06, "loss": 2.4333, "step": 920 }, { "epoch": 0.9006815968841285, "grad_norm": 1.6186519953923517, "learning_rate": 1.019890876883744e-06, "loss": 2.4319, "step": 925 }, { "epoch": 0.9055501460564752, "grad_norm": 1.586657649597322, "learning_rate": 1.0032887308245372e-06, "loss": 2.3888, "step": 930 }, { "epoch": 0.9104186952288218, "grad_norm": 1.6044497498274506, "learning_rate": 9.869453406918023e-07, "loss": 2.4039, "step": 935 }, { "epoch": 0.9152872444011685, "grad_norm": 1.6723989958086949, "learning_rate": 9.708570712098284e-07, "loss": 2.4005, "step": 940 }, { "epoch": 0.9201557935735151, "grad_norm": 1.6902859760264846, "learning_rate": 9.550203325714876e-07, "loss": 2.4261, "step": 945 }, { "epoch": 0.9250243427458618, "grad_norm": 1.6544740657432804, "learning_rate": 9.394315799398425e-07, "loss": 2.4451, "step": 950 }, { "epoch": 0.9298928919182083, "grad_norm": 1.54882787053935, "learning_rate": 9.240873129544315e-07, "loss": 2.444, "step": 955 }, { "epoch": 0.934761441090555, "grad_norm": 1.5909955762857377, "learning_rate": 9.089840752422235e-07, "loss": 2.3899, "step": 960 }, { "epoch": 0.9396299902629016, "grad_norm": 1.6145099358774662, "learning_rate": 8.941184539331783e-07, "loss": 2.4215, "step": 965 }, { "epoch": 0.9444985394352483, "grad_norm": 1.6027897997067373, "learning_rate": 8.794870791803967e-07, "loss": 2.4725, "step": 970 }, { "epoch": 0.9493670886075949, "grad_norm": 1.5457405929725805, "learning_rate": 8.650866236848001e-07, "loss": 2.4511, "step": 975 }, { "epoch": 0.9542356377799416, "grad_norm": 1.5741657202863206, "learning_rate": 8.509138022243367e-07, "loss": 2.4241, "step": 980 }, { "epoch": 0.9591041869522883, "grad_norm": 1.6377152121905334, "learning_rate": 8.369653711876414e-07, "loss": 2.4034, "step": 985 }, { "epoch": 0.9639727361246349, "grad_norm": 1.5672375026688792, "learning_rate": 8.232381281121472e-07, "loss": 2.4561, "step": 990 }, { "epoch": 0.9688412852969815, "grad_norm": 1.5788396321776008, "learning_rate": 8.097289112265904e-07, "loss": 2.4023, "step": 995 }, { "epoch": 0.9737098344693281, "grad_norm": 1.6492145516163694, "learning_rate": 7.964345989978902e-07, "loss": 2.4353, "step": 1000 }, { "epoch": 0.9737098344693281, "eval_loss": 2.4195876121520996, "eval_runtime": 85.322, "eval_samples_per_second": 85.547, "eval_steps_per_second": 0.68, "step": 1000 }, { "epoch": 0.9785783836416748, "grad_norm": 1.5879573832642313, "learning_rate": 7.83352109682346e-07, "loss": 2.4211, "step": 1005 }, { "epoch": 0.9834469328140214, "grad_norm": 1.6569952022837342, "learning_rate": 7.704784008811466e-07, "loss": 2.4405, "step": 1010 }, { "epoch": 0.9883154819863681, "grad_norm": 1.59377337317319, "learning_rate": 7.5781046910014e-07, "loss": 2.4239, "step": 1015 }, { "epoch": 0.9931840311587147, "grad_norm": 1.6314113718182768, "learning_rate": 7.453453493138208e-07, "loss": 2.3844, "step": 1020 }, { "epoch": 0.9980525803310614, "grad_norm": 1.6620011915571502, "learning_rate": 7.330801145335265e-07, "loss": 2.4181, "step": 1025 }, { "epoch": 1.002921129503408, "grad_norm": 1.9251273144439607, "learning_rate": 7.210118753797866e-07, "loss": 2.3064, "step": 1030 }, { "epoch": 1.0077896786757545, "grad_norm": 1.6577940198631569, "learning_rate": 7.091377796588075e-07, "loss": 2.1998, "step": 1035 }, { "epoch": 1.0126582278481013, "grad_norm": 1.7530256550139145, "learning_rate": 6.974550119430372e-07, "loss": 2.2708, "step": 1040 }, { "epoch": 1.017526777020448, "grad_norm": 1.6969458887579825, "learning_rate": 6.859607931558122e-07, "loss": 2.2173, "step": 1045 }, { "epoch": 1.0223953261927945, "grad_norm": 1.7016018631062932, "learning_rate": 6.746523801600123e-07, "loss": 2.2381, "step": 1050 }, { "epoch": 1.0272638753651413, "grad_norm": 1.7593644025263198, "learning_rate": 6.635270653507262e-07, "loss": 2.2734, "step": 1055 }, { "epoch": 1.0321324245374879, "grad_norm": 1.654744797257189, "learning_rate": 6.525821762518746e-07, "loss": 2.2064, "step": 1060 }, { "epoch": 1.0370009737098345, "grad_norm": 1.7421937503061258, "learning_rate": 6.41815075116768e-07, "loss": 2.2413, "step": 1065 }, { "epoch": 1.041869522882181, "grad_norm": 1.6658497077947585, "learning_rate": 6.312231585325566e-07, "loss": 2.2099, "step": 1070 }, { "epoch": 1.0467380720545278, "grad_norm": 1.6993685133537, "learning_rate": 6.208038570285602e-07, "loss": 2.2881, "step": 1075 }, { "epoch": 1.0516066212268744, "grad_norm": 1.7131084770608689, "learning_rate": 6.105546346884208e-07, "loss": 2.2381, "step": 1080 }, { "epoch": 1.056475170399221, "grad_norm": 1.7081524078028645, "learning_rate": 6.004729887660704e-07, "loss": 2.236, "step": 1085 }, { "epoch": 1.0613437195715676, "grad_norm": 1.8169379524340188, "learning_rate": 5.905564493054672e-07, "loss": 2.2342, "step": 1090 }, { "epoch": 1.0662122687439144, "grad_norm": 1.6685262696722296, "learning_rate": 5.808025787640756e-07, "loss": 2.222, "step": 1095 }, { "epoch": 1.071080817916261, "grad_norm": 1.6952684753454728, "learning_rate": 5.712089716400602e-07, "loss": 2.212, "step": 1100 }, { "epoch": 1.0759493670886076, "grad_norm": 1.7168463474300166, "learning_rate": 5.617732541031583e-07, "loss": 2.2186, "step": 1105 }, { "epoch": 1.0808179162609541, "grad_norm": 1.750097139848988, "learning_rate": 5.524930836292023e-07, "loss": 2.2494, "step": 1110 }, { "epoch": 1.085686465433301, "grad_norm": 1.7330864953187293, "learning_rate": 5.433661486382679e-07, "loss": 2.2102, "step": 1115 }, { "epoch": 1.0905550146056475, "grad_norm": 1.6803544506498977, "learning_rate": 5.343901681364102e-07, "loss": 2.2419, "step": 1120 }, { "epoch": 1.095423563777994, "grad_norm": 1.7013937838659918, "learning_rate": 5.255628913609601e-07, "loss": 2.2194, "step": 1125 }, { "epoch": 1.1002921129503407, "grad_norm": 1.7495270131848708, "learning_rate": 5.168820974293567e-07, "loss": 2.2213, "step": 1130 }, { "epoch": 1.1051606621226875, "grad_norm": 1.7709656106164202, "learning_rate": 5.08345594991474e-07, "loss": 2.2428, "step": 1135 }, { "epoch": 1.110029211295034, "grad_norm": 1.7338624116926167, "learning_rate": 4.999512218854316e-07, "loss": 2.1828, "step": 1140 }, { "epoch": 1.1148977604673806, "grad_norm": 1.6915851593817088, "learning_rate": 4.916968447968384e-07, "loss": 2.2286, "step": 1145 }, { "epoch": 1.1197663096397275, "grad_norm": 1.7318300441723782, "learning_rate": 4.835803589214607e-07, "loss": 2.2443, "step": 1150 }, { "epoch": 1.124634858812074, "grad_norm": 1.6773573534938775, "learning_rate": 4.7559968763127113e-07, "loss": 2.2427, "step": 1155 }, { "epoch": 1.1295034079844206, "grad_norm": 1.7187378671025355, "learning_rate": 4.67752782143862e-07, "loss": 2.2726, "step": 1160 }, { "epoch": 1.1343719571567672, "grad_norm": 1.740542786504217, "learning_rate": 4.6003762119518e-07, "loss": 2.1869, "step": 1165 }, { "epoch": 1.139240506329114, "grad_norm": 1.6945482053223642, "learning_rate": 4.524522107155746e-07, "loss": 2.2248, "step": 1170 }, { "epoch": 1.1441090555014606, "grad_norm": 1.6615466442429803, "learning_rate": 4.4499458350910965e-07, "loss": 2.2248, "step": 1175 }, { "epoch": 1.1489776046738072, "grad_norm": 1.7112548018169453, "learning_rate": 4.3766279893612986e-07, "loss": 2.2471, "step": 1180 }, { "epoch": 1.1538461538461537, "grad_norm": 1.860899105148827, "learning_rate": 4.304549425990434e-07, "loss": 2.2272, "step": 1185 }, { "epoch": 1.1587147030185005, "grad_norm": 1.691810639062161, "learning_rate": 4.23369126031296e-07, "loss": 2.263, "step": 1190 }, { "epoch": 1.1635832521908471, "grad_norm": 1.775044069455589, "learning_rate": 4.1640348638951344e-07, "loss": 2.2447, "step": 1195 }, { "epoch": 1.1684518013631937, "grad_norm": 1.9319359346716891, "learning_rate": 4.0955618614877875e-07, "loss": 2.2111, "step": 1200 }, { "epoch": 1.1684518013631937, "eval_loss": 2.429011106491089, "eval_runtime": 85.6795, "eval_samples_per_second": 85.19, "eval_steps_per_second": 0.677, "step": 1200 }, { "epoch": 1.1733203505355405, "grad_norm": 1.7574340980020409, "learning_rate": 4.0282541280102037e-07, "loss": 2.2332, "step": 1205 }, { "epoch": 1.178188899707887, "grad_norm": 1.6980825882291957, "learning_rate": 3.9620937855648926e-07, "loss": 2.2235, "step": 1210 }, { "epoch": 1.1830574488802337, "grad_norm": 1.7998015292888088, "learning_rate": 3.897063200482918e-07, "loss": 2.21, "step": 1215 }, { "epoch": 1.1879259980525803, "grad_norm": 1.631324517663901, "learning_rate": 3.833144980399548e-07, "loss": 2.219, "step": 1220 }, { "epoch": 1.192794547224927, "grad_norm": 1.6407460745986788, "learning_rate": 3.770321971360005e-07, "loss": 2.1846, "step": 1225 }, { "epoch": 1.1976630963972736, "grad_norm": 1.700809400503958, "learning_rate": 3.7085772549549546e-07, "loss": 2.2605, "step": 1230 }, { "epoch": 1.2025316455696202, "grad_norm": 1.6715825199948962, "learning_rate": 3.647894145485631e-07, "loss": 2.2213, "step": 1235 }, { "epoch": 1.2074001947419668, "grad_norm": 1.6992864948362203, "learning_rate": 3.588256187158179e-07, "loss": 2.2256, "step": 1240 }, { "epoch": 1.2122687439143136, "grad_norm": 1.7332695276540628, "learning_rate": 3.5296471513070895e-07, "loss": 2.2602, "step": 1245 }, { "epoch": 1.2171372930866602, "grad_norm": 1.7489006914371872, "learning_rate": 3.472051033647415e-07, "loss": 2.2514, "step": 1250 }, { "epoch": 1.2220058422590068, "grad_norm": 1.7282302399042118, "learning_rate": 3.4154520515555147e-07, "loss": 2.2403, "step": 1255 }, { "epoch": 1.2268743914313536, "grad_norm": 1.702612549564712, "learning_rate": 3.3598346413781214e-07, "loss": 2.2181, "step": 1260 }, { "epoch": 1.2317429406037002, "grad_norm": 1.7033226596241964, "learning_rate": 3.305183455769435e-07, "loss": 2.2761, "step": 1265 }, { "epoch": 1.2366114897760467, "grad_norm": 1.750385373058268, "learning_rate": 3.251483361056018e-07, "loss": 2.2387, "step": 1270 }, { "epoch": 1.2414800389483933, "grad_norm": 1.7510710962758804, "learning_rate": 3.1987194346292783e-07, "loss": 2.2642, "step": 1275 }, { "epoch": 1.24634858812074, "grad_norm": 1.699193378636474, "learning_rate": 3.1468769623652614e-07, "loss": 2.2675, "step": 1280 }, { "epoch": 1.2512171372930867, "grad_norm": 1.6670573040794279, "learning_rate": 3.0959414360715174e-07, "loss": 2.2596, "step": 1285 }, { "epoch": 1.2560856864654333, "grad_norm": 1.7667142100191355, "learning_rate": 3.045898550960847e-07, "loss": 2.2037, "step": 1290 }, { "epoch": 1.2609542356377799, "grad_norm": 1.7069461836357607, "learning_rate": 2.9967342031515984e-07, "loss": 2.1602, "step": 1295 }, { "epoch": 1.2658227848101267, "grad_norm": 1.6514498602889256, "learning_rate": 2.948434487194415e-07, "loss": 2.2278, "step": 1300 }, { "epoch": 1.2706913339824732, "grad_norm": 1.661661974641321, "learning_rate": 2.900985693625059e-07, "loss": 2.2009, "step": 1305 }, { "epoch": 1.2755598831548198, "grad_norm": 1.774692203861925, "learning_rate": 2.854374306543197e-07, "loss": 2.2549, "step": 1310 }, { "epoch": 1.2804284323271666, "grad_norm": 1.678761386598376, "learning_rate": 2.8085870012168447e-07, "loss": 2.2362, "step": 1315 }, { "epoch": 1.2852969814995132, "grad_norm": 1.717206031840906, "learning_rate": 2.763610641712288e-07, "loss": 2.2259, "step": 1320 }, { "epoch": 1.2901655306718598, "grad_norm": 1.725788802654923, "learning_rate": 2.7194322785492024e-07, "loss": 2.2743, "step": 1325 }, { "epoch": 1.2950340798442064, "grad_norm": 1.7272128729108773, "learning_rate": 2.6760391463808273e-07, "loss": 2.2238, "step": 1330 }, { "epoch": 1.299902629016553, "grad_norm": 1.673730997692294, "learning_rate": 2.633418661698872e-07, "loss": 2.213, "step": 1335 }, { "epoch": 1.3047711781888998, "grad_norm": 1.67421773725042, "learning_rate": 2.591558420563027e-07, "loss": 2.2159, "step": 1340 }, { "epoch": 1.3096397273612463, "grad_norm": 1.7181866282412452, "learning_rate": 2.550446196354801e-07, "loss": 2.1931, "step": 1345 }, { "epoch": 1.314508276533593, "grad_norm": 1.9995904186048974, "learning_rate": 2.5100699375554764e-07, "loss": 2.2394, "step": 1350 }, { "epoch": 1.3193768257059397, "grad_norm": 1.6972697466343885, "learning_rate": 2.4704177655480044e-07, "loss": 2.2046, "step": 1355 }, { "epoch": 1.3242453748782863, "grad_norm": 1.7352888329956384, "learning_rate": 2.431477972442557e-07, "loss": 2.2582, "step": 1360 }, { "epoch": 1.3291139240506329, "grad_norm": 1.7536755936521693, "learning_rate": 2.3932390189255786e-07, "loss": 2.2573, "step": 1365 }, { "epoch": 1.3339824732229795, "grad_norm": 1.7063896078961662, "learning_rate": 2.3556895321321132e-07, "loss": 2.2665, "step": 1370 }, { "epoch": 1.3388510223953263, "grad_norm": 1.6740223494854998, "learning_rate": 2.3188183035411849e-07, "loss": 2.2163, "step": 1375 }, { "epoch": 1.3437195715676729, "grad_norm": 1.7492684308128166, "learning_rate": 2.2826142868940236e-07, "loss": 2.1842, "step": 1380 }, { "epoch": 1.3485881207400194, "grad_norm": 1.7405438484291114, "learning_rate": 2.2470665961349557e-07, "loss": 2.1983, "step": 1385 }, { "epoch": 1.353456669912366, "grad_norm": 1.764967629336457, "learning_rate": 2.2121645033746942e-07, "loss": 2.2207, "step": 1390 }, { "epoch": 1.3583252190847128, "grad_norm": 1.7077892049631243, "learning_rate": 2.1778974368759137e-07, "loss": 2.2475, "step": 1395 }, { "epoch": 1.3631937682570594, "grad_norm": 1.7009940340048213, "learning_rate": 2.1442549790608077e-07, "loss": 2.2503, "step": 1400 }, { "epoch": 1.3631937682570594, "eval_loss": 2.428063154220581, "eval_runtime": 85.6482, "eval_samples_per_second": 85.221, "eval_steps_per_second": 0.677, "step": 1400 }, { "epoch": 1.368062317429406, "grad_norm": 1.6852036518237539, "learning_rate": 2.1112268645405111e-07, "loss": 2.2092, "step": 1405 }, { "epoch": 1.3729308666017528, "grad_norm": 1.7762881669503894, "learning_rate": 2.0788029781661377e-07, "loss": 2.275, "step": 1410 }, { "epoch": 1.3777994157740994, "grad_norm": 1.7187728321020497, "learning_rate": 2.0469733531012517e-07, "loss": 2.205, "step": 1415 }, { "epoch": 1.382667964946446, "grad_norm": 1.75561614859316, "learning_rate": 2.015728168915585e-07, "loss": 2.2513, "step": 1420 }, { "epoch": 1.3875365141187925, "grad_norm": 1.7108273702116936, "learning_rate": 1.9850577496997802e-07, "loss": 2.3129, "step": 1425 }, { "epoch": 1.3924050632911391, "grad_norm": 1.7373174761439103, "learning_rate": 1.9549525622009822e-07, "loss": 2.3009, "step": 1430 }, { "epoch": 1.397273612463486, "grad_norm": 1.6944981878481962, "learning_rate": 1.9254032139790948e-07, "loss": 2.2276, "step": 1435 }, { "epoch": 1.4021421616358325, "grad_norm": 1.7020044804533232, "learning_rate": 1.8964004515834934e-07, "loss": 2.2515, "step": 1440 }, { "epoch": 1.407010710808179, "grad_norm": 1.7104910024357038, "learning_rate": 1.8679351587500074e-07, "loss": 2.2176, "step": 1445 }, { "epoch": 1.4118792599805259, "grad_norm": 1.706154278461277, "learning_rate": 1.839998354618001e-07, "loss": 2.2426, "step": 1450 }, { "epoch": 1.4167478091528725, "grad_norm": 1.6546136448975106, "learning_rate": 1.8125811919673288e-07, "loss": 2.2729, "step": 1455 }, { "epoch": 1.421616358325219, "grad_norm": 1.7298929182929772, "learning_rate": 1.7856749554750352e-07, "loss": 2.2213, "step": 1460 }, { "epoch": 1.4264849074975658, "grad_norm": 1.6799052546726045, "learning_rate": 1.7592710599915514e-07, "loss": 2.2494, "step": 1465 }, { "epoch": 1.4313534566699124, "grad_norm": 1.7192873393376786, "learning_rate": 1.7333610488362632e-07, "loss": 2.2219, "step": 1470 }, { "epoch": 1.436222005842259, "grad_norm": 1.6974268468535265, "learning_rate": 1.7079365921122288e-07, "loss": 2.2159, "step": 1475 }, { "epoch": 1.4410905550146056, "grad_norm": 1.736932687979258, "learning_rate": 1.682989485039898e-07, "loss": 2.2001, "step": 1480 }, { "epoch": 1.4459591041869522, "grad_norm": 1.7389015197229705, "learning_rate": 1.65851164630961e-07, "loss": 2.1815, "step": 1485 }, { "epoch": 1.450827653359299, "grad_norm": 1.7402452727637348, "learning_rate": 1.634495116452756e-07, "loss": 2.274, "step": 1490 }, { "epoch": 1.4556962025316456, "grad_norm": 1.689117128248476, "learning_rate": 1.6109320562313633e-07, "loss": 2.2296, "step": 1495 }, { "epoch": 1.4605647517039921, "grad_norm": 1.7021757020476915, "learning_rate": 1.587814745045978e-07, "loss": 2.2488, "step": 1500 }, { "epoch": 1.465433300876339, "grad_norm": 1.726585875668609, "learning_rate": 1.565135579361648e-07, "loss": 2.248, "step": 1505 }, { "epoch": 1.4703018500486855, "grad_norm": 1.7896520417513544, "learning_rate": 1.542887071151836e-07, "loss": 2.2768, "step": 1510 }, { "epoch": 1.475170399221032, "grad_norm": 1.8424772477578564, "learning_rate": 1.5210618463601055e-07, "loss": 2.2177, "step": 1515 }, { "epoch": 1.480038948393379, "grad_norm": 1.7094835426748392, "learning_rate": 1.4996526433793895e-07, "loss": 2.1814, "step": 1520 }, { "epoch": 1.4849074975657255, "grad_norm": 1.774002445036824, "learning_rate": 1.478652311548687e-07, "loss": 2.241, "step": 1525 }, { "epoch": 1.489776046738072, "grad_norm": 1.7528924437568645, "learning_rate": 1.4580538096670264e-07, "loss": 2.2446, "step": 1530 }, { "epoch": 1.4946445959104186, "grad_norm": 1.7450272228199162, "learning_rate": 1.4378502045245165e-07, "loss": 2.2098, "step": 1535 }, { "epoch": 1.4995131450827652, "grad_norm": 1.71350388367754, "learning_rate": 1.4180346694503253e-07, "loss": 2.2407, "step": 1540 }, { "epoch": 1.504381694255112, "grad_norm": 1.6655187619092102, "learning_rate": 1.398600482877437e-07, "loss": 2.2516, "step": 1545 }, { "epoch": 1.5092502434274586, "grad_norm": 1.730038860188208, "learning_rate": 1.379541026923992e-07, "loss": 2.2244, "step": 1550 }, { "epoch": 1.5141187925998052, "grad_norm": 1.7106880841934986, "learning_rate": 1.360849785991099e-07, "loss": 2.2359, "step": 1555 }, { "epoch": 1.518987341772152, "grad_norm": 1.7507950020460012, "learning_rate": 1.3425203453769053e-07, "loss": 2.2067, "step": 1560 }, { "epoch": 1.5238558909444986, "grad_norm": 1.6877714873976932, "learning_rate": 1.3245463899068173e-07, "loss": 2.2203, "step": 1565 }, { "epoch": 1.5287244401168452, "grad_norm": 1.7990198522805025, "learning_rate": 1.306921702579676e-07, "loss": 2.2176, "step": 1570 }, { "epoch": 1.533592989289192, "grad_norm": 1.7072472713071833, "learning_rate": 1.2896401632297622e-07, "loss": 2.2199, "step": 1575 }, { "epoch": 1.5384615384615383, "grad_norm": 1.7137109372098451, "learning_rate": 1.2726957472044431e-07, "loss": 2.2593, "step": 1580 }, { "epoch": 1.5433300876338851, "grad_norm": 1.7487536889121376, "learning_rate": 1.2560825240573496e-07, "loss": 2.2512, "step": 1585 }, { "epoch": 1.5481986368062317, "grad_norm": 1.689517701653058, "learning_rate": 1.23979465625689e-07, "loss": 2.2594, "step": 1590 }, { "epoch": 1.5530671859785783, "grad_norm": 1.7198840121589114, "learning_rate": 1.223826397909982e-07, "loss": 2.1724, "step": 1595 }, { "epoch": 1.557935735150925, "grad_norm": 1.7280648830031222, "learning_rate": 1.2081720935008362e-07, "loss": 2.258, "step": 1600 }, { "epoch": 1.557935735150925, "eval_loss": 2.4271326065063477, "eval_runtime": 85.3342, "eval_samples_per_second": 85.534, "eval_steps_per_second": 0.68, "step": 1600 }, { "epoch": 1.5628042843232717, "grad_norm": 1.7166373586948793, "learning_rate": 1.1928261766446461e-07, "loss": 2.2507, "step": 1605 }, { "epoch": 1.5676728334956183, "grad_norm": 1.6962264139269825, "learning_rate": 1.1777831688560428e-07, "loss": 2.214, "step": 1610 }, { "epoch": 1.572541382667965, "grad_norm": 1.667890993114025, "learning_rate": 1.1630376783321605e-07, "loss": 2.2107, "step": 1615 }, { "epoch": 1.5774099318403116, "grad_norm": 1.7705130739867874, "learning_rate": 1.1485843987501698e-07, "loss": 2.2298, "step": 1620 }, { "epoch": 1.5822784810126582, "grad_norm": 1.7067117715143372, "learning_rate": 1.134418108079144e-07, "loss": 2.2568, "step": 1625 }, { "epoch": 1.587147030185005, "grad_norm": 1.698686461270445, "learning_rate": 1.1205336674061039e-07, "loss": 2.256, "step": 1630 }, { "epoch": 1.5920155793573514, "grad_norm": 1.6914580859058206, "learning_rate": 1.1069260197761055e-07, "loss": 2.2255, "step": 1635 }, { "epoch": 1.5968841285296982, "grad_norm": 1.695162008456652, "learning_rate": 1.0935901890462346e-07, "loss": 2.1925, "step": 1640 }, { "epoch": 1.6017526777020448, "grad_norm": 1.8353600777066885, "learning_rate": 1.0805212787533523e-07, "loss": 2.259, "step": 1645 }, { "epoch": 1.6066212268743914, "grad_norm": 1.7146893423140706, "learning_rate": 1.0677144709954816e-07, "loss": 2.251, "step": 1650 }, { "epoch": 1.6114897760467382, "grad_norm": 1.7479655666276084, "learning_rate": 1.0551650253266659e-07, "loss": 2.2297, "step": 1655 }, { "epoch": 1.6163583252190847, "grad_norm": 1.7879147626083056, "learning_rate": 1.0428682776651918e-07, "loss": 2.1994, "step": 1660 }, { "epoch": 1.6212268743914313, "grad_norm": 1.7970614657397905, "learning_rate": 1.030819639215023e-07, "loss": 2.2049, "step": 1665 }, { "epoch": 1.6260954235637781, "grad_norm": 1.7518141901380913, "learning_rate": 1.0190145954003192e-07, "loss": 2.2203, "step": 1670 }, { "epoch": 1.6309639727361245, "grad_norm": 1.7501918042535358, "learning_rate": 1.0074487048129093e-07, "loss": 2.2585, "step": 1675 }, { "epoch": 1.6358325219084713, "grad_norm": 1.756138340727838, "learning_rate": 9.961175981725802e-08, "loss": 2.2441, "step": 1680 }, { "epoch": 1.6407010710808179, "grad_norm": 1.6916112013631275, "learning_rate": 9.850169773000545e-08, "loss": 2.2229, "step": 1685 }, { "epoch": 1.6455696202531644, "grad_norm": 1.7520093451689733, "learning_rate": 9.741426141025332e-08, "loss": 2.2418, "step": 1690 }, { "epoch": 1.6504381694255112, "grad_norm": 1.721678894466571, "learning_rate": 9.634903495716675e-08, "loss": 2.215, "step": 1695 }, { "epoch": 1.6553067185978578, "grad_norm": 1.7726004550340564, "learning_rate": 9.530560927938334e-08, "loss": 2.2609, "step": 1700 }, { "epoch": 1.6601752677702044, "grad_norm": 1.7520211987580594, "learning_rate": 9.428358199725911e-08, "loss": 2.2388, "step": 1705 }, { "epoch": 1.6650438169425512, "grad_norm": 1.8152868603337584, "learning_rate": 9.328255734631862e-08, "loss": 2.2346, "step": 1710 }, { "epoch": 1.6699123661148978, "grad_norm": 1.7051369362258122, "learning_rate": 9.230214608189946e-08, "loss": 2.201, "step": 1715 }, { "epoch": 1.6747809152872444, "grad_norm": 1.7253608618380802, "learning_rate": 9.1341965384976e-08, "loss": 2.2579, "step": 1720 }, { "epoch": 1.6796494644595912, "grad_norm": 1.6955731274986756, "learning_rate": 9.040163876915256e-08, "loss": 2.2456, "step": 1725 }, { "epoch": 1.6845180136319375, "grad_norm": 1.8315504305607004, "learning_rate": 8.948079598881221e-08, "loss": 2.2085, "step": 1730 }, { "epoch": 1.6893865628042843, "grad_norm": 1.680264316266455, "learning_rate": 8.857907294841052e-08, "loss": 2.2427, "step": 1735 }, { "epoch": 1.694255111976631, "grad_norm": 1.7344138245065615, "learning_rate": 8.76961116129008e-08, "loss": 2.2503, "step": 1740 }, { "epoch": 1.6991236611489775, "grad_norm": 1.7400368127622114, "learning_rate": 8.683155991928096e-08, "loss": 2.2245, "step": 1745 }, { "epoch": 1.7039922103213243, "grad_norm": 1.7436841437009907, "learning_rate": 8.598507168924832e-08, "loss": 2.229, "step": 1750 }, { "epoch": 1.7088607594936709, "grad_norm": 1.718735182879372, "learning_rate": 8.515630654295236e-08, "loss": 2.214, "step": 1755 }, { "epoch": 1.7137293086660175, "grad_norm": 2.4325656334712416, "learning_rate": 8.43449298138329e-08, "loss": 2.2273, "step": 1760 }, { "epoch": 1.7185978578383643, "grad_norm": 1.898796758188917, "learning_rate": 8.355061246453264e-08, "loss": 2.2182, "step": 1765 }, { "epoch": 1.7234664070107109, "grad_norm": 1.7705269671220705, "learning_rate": 8.277303100387306e-08, "loss": 2.2375, "step": 1770 }, { "epoch": 1.7283349561830574, "grad_norm": 1.7467032834457301, "learning_rate": 8.201186740488162e-08, "loss": 2.2547, "step": 1775 }, { "epoch": 1.7332035053554042, "grad_norm": 1.7519508473607348, "learning_rate": 8.126680902385984e-08, "loss": 2.2533, "step": 1780 }, { "epoch": 1.7380720545277506, "grad_norm": 1.6602010890382983, "learning_rate": 8.053754852048118e-08, "loss": 2.2409, "step": 1785 }, { "epoch": 1.7429406037000974, "grad_norm": 1.7182534815184765, "learning_rate": 7.982378377890754e-08, "loss": 2.2169, "step": 1790 }, { "epoch": 1.747809152872444, "grad_norm": 1.7132799050190683, "learning_rate": 7.912521782991345e-08, "loss": 2.21, "step": 1795 }, { "epoch": 1.7526777020447906, "grad_norm": 1.7492486479656935, "learning_rate": 7.844155877400776e-08, "loss": 2.254, "step": 1800 }, { "epoch": 1.7526777020447906, "eval_loss": 2.4266231060028076, "eval_runtime": 84.8728, "eval_samples_per_second": 85.999, "eval_steps_per_second": 0.683, "step": 1800 }, { "epoch": 1.7575462512171374, "grad_norm": 1.7982513495954076, "learning_rate": 7.777251970554109e-08, "loss": 2.2446, "step": 1805 }, { "epoch": 1.762414800389484, "grad_norm": 1.7235517571422887, "learning_rate": 7.711781863779007e-08, "loss": 2.1939, "step": 1810 }, { "epoch": 1.7672833495618305, "grad_norm": 1.7506770945570296, "learning_rate": 7.64771784290061e-08, "loss": 2.2339, "step": 1815 }, { "epoch": 1.7721518987341773, "grad_norm": 1.961410951418005, "learning_rate": 7.58503267094196e-08, "loss": 2.2686, "step": 1820 }, { "epoch": 1.7770204479065237, "grad_norm": 1.8140605591823493, "learning_rate": 7.523699580918899e-08, "loss": 2.2136, "step": 1825 }, { "epoch": 1.7818889970788705, "grad_norm": 1.733018609908853, "learning_rate": 7.463692268728381e-08, "loss": 2.165, "step": 1830 }, { "epoch": 1.7867575462512173, "grad_norm": 1.6824331979801876, "learning_rate": 7.404984886129268e-08, "loss": 2.2369, "step": 1835 }, { "epoch": 1.7916260954235637, "grad_norm": 1.8520431760674112, "learning_rate": 7.347552033814525e-08, "loss": 2.2612, "step": 1840 }, { "epoch": 1.7964946445959105, "grad_norm": 1.7117045631055423, "learning_rate": 7.291368754573846e-08, "loss": 2.2425, "step": 1845 }, { "epoch": 1.801363193768257, "grad_norm": 1.7052819169533668, "learning_rate": 7.236410526545777e-08, "loss": 2.196, "step": 1850 }, { "epoch": 1.8062317429406036, "grad_norm": 1.6384344799557868, "learning_rate": 7.182653256558277e-08, "loss": 2.2563, "step": 1855 }, { "epoch": 1.8111002921129504, "grad_norm": 1.726079611316107, "learning_rate": 7.130073273556794e-08, "loss": 2.2455, "step": 1860 }, { "epoch": 1.815968841285297, "grad_norm": 1.8072162191497745, "learning_rate": 7.07864732211891e-08, "loss": 2.2181, "step": 1865 }, { "epoch": 1.8208373904576436, "grad_norm": 1.720087653439122, "learning_rate": 7.028352556054533e-08, "loss": 2.2295, "step": 1870 }, { "epoch": 1.8257059396299904, "grad_norm": 1.7654960116615692, "learning_rate": 6.979166532090796e-08, "loss": 2.2391, "step": 1875 }, { "epoch": 1.8305744888023368, "grad_norm": 1.735053092580652, "learning_rate": 6.931067203640622e-08, "loss": 2.2372, "step": 1880 }, { "epoch": 1.8354430379746836, "grad_norm": 0.8624508534667337, "learning_rate": 6.884032914654115e-08, "loss": 2.2893, "step": 1885 }, { "epoch": 1.8403115871470301, "grad_norm": 1.7141597133248032, "learning_rate": 6.838042393551797e-08, "loss": 2.1609, "step": 1890 }, { "epoch": 1.8451801363193767, "grad_norm": 1.7126007991228827, "learning_rate": 6.793074747238838e-08, "loss": 2.1911, "step": 1895 }, { "epoch": 1.8500486854917235, "grad_norm": 1.7133412238684549, "learning_rate": 6.749109455199282e-08, "loss": 2.2345, "step": 1900 }, { "epoch": 1.85491723466407, "grad_norm": 1.7274466496979972, "learning_rate": 6.706126363669515e-08, "loss": 2.2514, "step": 1905 }, { "epoch": 1.8597857838364167, "grad_norm": 1.7189469301901343, "learning_rate": 6.664105679889947e-08, "loss": 2.2458, "step": 1910 }, { "epoch": 1.8646543330087635, "grad_norm": 1.690182741334165, "learning_rate": 6.623027966434113e-08, "loss": 2.2038, "step": 1915 }, { "epoch": 1.86952288218111, "grad_norm": 1.7617528535634877, "learning_rate": 6.582874135614325e-08, "loss": 2.1569, "step": 1920 }, { "epoch": 1.8743914313534566, "grad_norm": 1.7229096608807917, "learning_rate": 6.54362544396295e-08, "loss": 2.2416, "step": 1925 }, { "epoch": 1.8792599805258035, "grad_norm": 1.796873830830146, "learning_rate": 6.505263486788538e-08, "loss": 2.2388, "step": 1930 }, { "epoch": 1.8841285296981498, "grad_norm": 1.6550586704575283, "learning_rate": 6.46777019280588e-08, "loss": 2.2445, "step": 1935 }, { "epoch": 1.8889970788704966, "grad_norm": 1.7133733906940865, "learning_rate": 6.431127818839212e-08, "loss": 2.1907, "step": 1940 }, { "epoch": 1.8938656280428432, "grad_norm": 1.7131092326344233, "learning_rate": 6.395318944597702e-08, "loss": 2.2508, "step": 1945 }, { "epoch": 1.8987341772151898, "grad_norm": 1.7975356797064426, "learning_rate": 6.360326467522418e-08, "loss": 2.2619, "step": 1950 }, { "epoch": 1.9036027263875366, "grad_norm": 1.7927617071731754, "learning_rate": 6.326133597703928e-08, "loss": 2.2306, "step": 1955 }, { "epoch": 1.9084712755598832, "grad_norm": 1.7327613795288057, "learning_rate": 6.292723852869772e-08, "loss": 2.2258, "step": 1960 }, { "epoch": 1.9133398247322297, "grad_norm": 1.7114888287588472, "learning_rate": 6.260081053440938e-08, "loss": 2.2258, "step": 1965 }, { "epoch": 1.9182083739045765, "grad_norm": 1.7617910174679117, "learning_rate": 6.228189317656625e-08, "loss": 2.2214, "step": 1970 }, { "epoch": 1.9230769230769231, "grad_norm": 1.69813136105815, "learning_rate": 6.197033056766434e-08, "loss": 2.2189, "step": 1975 }, { "epoch": 1.9279454722492697, "grad_norm": 1.668741197040639, "learning_rate": 6.166596970289254e-08, "loss": 2.2168, "step": 1980 }, { "epoch": 1.9328140214216165, "grad_norm": 1.7177940007984969, "learning_rate": 6.136866041338061e-08, "loss": 2.2131, "step": 1985 }, { "epoch": 1.9376825705939629, "grad_norm": 2.097413251733257, "learning_rate": 6.107825532009846e-08, "loss": 2.2431, "step": 1990 }, { "epoch": 1.9425511197663097, "grad_norm": 1.6798259076762905, "learning_rate": 6.079460978839945e-08, "loss": 2.273, "step": 1995 }, { "epoch": 1.9474196689386563, "grad_norm": 1.7260224966683506, "learning_rate": 6.051758188319985e-08, "loss": 2.2508, "step": 2000 }, { "epoch": 1.9474196689386563, "eval_loss": 2.4265518188476562, "eval_runtime": 85.2652, "eval_samples_per_second": 85.604, "eval_steps_per_second": 0.68, "step": 2000 }, { "epoch": 1.9522882181110028, "grad_norm": 1.6953303266550364, "learning_rate": 6.024703232478723e-08, "loss": 2.2536, "step": 2005 }, { "epoch": 1.9571567672833496, "grad_norm": 1.7537091815900079, "learning_rate": 5.998282444525052e-08, "loss": 2.1932, "step": 2010 }, { "epoch": 1.9620253164556962, "grad_norm": 1.6913312559085918, "learning_rate": 5.97248241455241e-08, "loss": 2.2097, "step": 2015 }, { "epoch": 1.9668938656280428, "grad_norm": 1.709640143824205, "learning_rate": 5.947289985303898e-08, "loss": 2.2739, "step": 2020 }, { "epoch": 1.9717624148003896, "grad_norm": 1.769591500268775, "learning_rate": 5.922692247997387e-08, "loss": 2.2345, "step": 2025 }, { "epoch": 1.976630963972736, "grad_norm": 1.7089545500913113, "learning_rate": 5.898676538209882e-08, "loss": 2.1989, "step": 2030 }, { "epoch": 1.9814995131450828, "grad_norm": 1.7324755039666533, "learning_rate": 5.8752304318204655e-08, "loss": 2.2296, "step": 2035 }, { "epoch": 1.9863680623174296, "grad_norm": 1.7254800842506772, "learning_rate": 5.852341741011111e-08, "loss": 2.193, "step": 2040 }, { "epoch": 1.991236611489776, "grad_norm": 1.7451878538885703, "learning_rate": 5.829998510324686e-08, "loss": 2.2829, "step": 2045 }, { "epoch": 1.9961051606621227, "grad_norm": 1.7607671547384287, "learning_rate": 5.8081890127794496e-08, "loss": 2.2133, "step": 2050 }, { "epoch": 2.0009737098344695, "grad_norm": 1.7100416555620679, "learning_rate": 5.7869017460393855e-08, "loss": 2.233, "step": 2055 }, { "epoch": 2.005842259006816, "grad_norm": 1.746899480839477, "learning_rate": 5.7661254286396653e-08, "loss": 2.2424, "step": 2060 }, { "epoch": 2.0107108081791627, "grad_norm": 1.7305460820757212, "learning_rate": 5.74584899626664e-08, "loss": 2.1762, "step": 2065 }, { "epoch": 2.015579357351509, "grad_norm": 1.7063747134288176, "learning_rate": 5.726061598091625e-08, "loss": 2.2345, "step": 2070 }, { "epoch": 2.020447906523856, "grad_norm": 1.7115321830413128, "learning_rate": 5.706752593157899e-08, "loss": 2.2415, "step": 2075 }, { "epoch": 2.0253164556962027, "grad_norm": 1.7315759852916324, "learning_rate": 5.687911546820225e-08, "loss": 2.2013, "step": 2080 }, { "epoch": 2.030185004868549, "grad_norm": 1.7053153190560344, "learning_rate": 5.669528227236286e-08, "loss": 2.1833, "step": 2085 }, { "epoch": 2.035053554040896, "grad_norm": 1.7480977685627526, "learning_rate": 5.6515926019093655e-08, "loss": 2.211, "step": 2090 }, { "epoch": 2.0399221032132426, "grad_norm": 1.7325131107565894, "learning_rate": 5.6340948342816956e-08, "loss": 2.1746, "step": 2095 }, { "epoch": 2.044790652385589, "grad_norm": 1.7861322062259097, "learning_rate": 5.617025280377817e-08, "loss": 2.2137, "step": 2100 }, { "epoch": 2.049659201557936, "grad_norm": 1.7407845229020498, "learning_rate": 5.6003744854973425e-08, "loss": 2.2327, "step": 2105 }, { "epoch": 2.0545277507302826, "grad_norm": 1.7703593488238105, "learning_rate": 5.584133180956534e-08, "loss": 2.2133, "step": 2110 }, { "epoch": 2.059396299902629, "grad_norm": 1.7217536220585616, "learning_rate": 5.568292280878073e-08, "loss": 2.1892, "step": 2115 }, { "epoch": 2.0642648490749758, "grad_norm": 1.7276020192361135, "learning_rate": 5.552842879028437e-08, "loss": 2.2204, "step": 2120 }, { "epoch": 2.069133398247322, "grad_norm": 1.727960780515485, "learning_rate": 5.537776245702285e-08, "loss": 2.1996, "step": 2125 }, { "epoch": 2.074001947419669, "grad_norm": 1.7514773544281632, "learning_rate": 5.523083824653292e-08, "loss": 2.1812, "step": 2130 }, { "epoch": 2.0788704965920157, "grad_norm": 1.7204243933149874, "learning_rate": 5.5087572300708064e-08, "loss": 2.2412, "step": 2135 }, { "epoch": 2.083739045764362, "grad_norm": 1.7770689913358608, "learning_rate": 5.49478824360182e-08, "loss": 2.2175, "step": 2140 }, { "epoch": 2.088607594936709, "grad_norm": 1.8118845495638691, "learning_rate": 5.4811688114176284e-08, "loss": 2.2016, "step": 2145 }, { "epoch": 2.0934761441090557, "grad_norm": 1.746154499517824, "learning_rate": 5.46789104132466e-08, "loss": 2.2338, "step": 2150 }, { "epoch": 2.098344693281402, "grad_norm": 1.7040405445808213, "learning_rate": 5.454947199918886e-08, "loss": 2.2298, "step": 2155 }, { "epoch": 2.103213242453749, "grad_norm": 1.6905682803922102, "learning_rate": 5.4423297097832996e-08, "loss": 2.2539, "step": 2160 }, { "epoch": 2.108081791626095, "grad_norm": 1.7302575758498189, "learning_rate": 5.430031146727882e-08, "loss": 2.2279, "step": 2165 }, { "epoch": 2.112950340798442, "grad_norm": 1.771661817904592, "learning_rate": 5.418044237071543e-08, "loss": 2.1828, "step": 2170 }, { "epoch": 2.117818889970789, "grad_norm": 1.726387016486399, "learning_rate": 5.406361854965489e-08, "loss": 2.2167, "step": 2175 }, { "epoch": 2.122687439143135, "grad_norm": 1.72736148684106, "learning_rate": 5.394977019757503e-08, "loss": 2.1627, "step": 2180 }, { "epoch": 2.127555988315482, "grad_norm": 1.732688655222483, "learning_rate": 5.3838828933965965e-08, "loss": 2.2482, "step": 2185 }, { "epoch": 2.132424537487829, "grad_norm": 1.765418024490754, "learning_rate": 5.373072777877539e-08, "loss": 2.2416, "step": 2190 }, { "epoch": 2.137293086660175, "grad_norm": 1.7905409309741185, "learning_rate": 5.3625401127247335e-08, "loss": 2.2049, "step": 2195 }, { "epoch": 2.142161635832522, "grad_norm": 1.7285450482838325, "learning_rate": 5.3522784725149425e-08, "loss": 2.2112, "step": 2200 }, { "epoch": 2.142161635832522, "eval_loss": 2.4287049770355225, "eval_runtime": 85.2698, "eval_samples_per_second": 85.599, "eval_steps_per_second": 0.68, "step": 2200 }, { "epoch": 2.1470301850048688, "grad_norm": 1.7298233620729722, "learning_rate": 5.3422815644383576e-08, "loss": 2.2268, "step": 2205 }, { "epoch": 2.151898734177215, "grad_norm": 1.7106243642725834, "learning_rate": 5.332543225897528e-08, "loss": 2.1823, "step": 2210 }, { "epoch": 2.156767283349562, "grad_norm": 1.7093421731817005, "learning_rate": 5.3230574221436374e-08, "loss": 2.2282, "step": 2215 }, { "epoch": 2.1616358325219083, "grad_norm": 1.7311568429821071, "learning_rate": 5.313818243949664e-08, "loss": 2.2213, "step": 2220 }, { "epoch": 2.166504381694255, "grad_norm": 1.7674044638597066, "learning_rate": 5.3048199053199294e-08, "loss": 2.2254, "step": 2225 }, { "epoch": 2.171372930866602, "grad_norm": 1.7115381111757135, "learning_rate": 5.296056741235573e-08, "loss": 2.2418, "step": 2230 }, { "epoch": 2.1762414800389482, "grad_norm": 1.6790013492814684, "learning_rate": 5.2875232054354564e-08, "loss": 2.2243, "step": 2235 }, { "epoch": 2.181110029211295, "grad_norm": 1.7513052402271196, "learning_rate": 5.279213868232073e-08, "loss": 2.247, "step": 2240 }, { "epoch": 2.185978578383642, "grad_norm": 1.8041144340469755, "learning_rate": 5.271123414361961e-08, "loss": 2.2318, "step": 2245 }, { "epoch": 2.190847127555988, "grad_norm": 1.711266621732132, "learning_rate": 5.263246640870184e-08, "loss": 2.2393, "step": 2250 }, { "epoch": 2.195715676728335, "grad_norm": 1.6969264596911122, "learning_rate": 5.2555784550284366e-08, "loss": 2.2087, "step": 2255 }, { "epoch": 2.2005842259006814, "grad_norm": 1.7651857140540494, "learning_rate": 5.2481138722863035e-08, "loss": 2.2069, "step": 2260 }, { "epoch": 2.205452775073028, "grad_norm": 1.7817072494501394, "learning_rate": 5.240848014255253e-08, "loss": 2.2051, "step": 2265 }, { "epoch": 2.210321324245375, "grad_norm": 1.7343663502264346, "learning_rate": 5.233776106724918e-08, "loss": 2.2389, "step": 2270 }, { "epoch": 2.2151898734177213, "grad_norm": 1.6921255572364247, "learning_rate": 5.2268934777112365e-08, "loss": 2.2362, "step": 2275 }, { "epoch": 2.220058422590068, "grad_norm": 1.7013539832560813, "learning_rate": 5.220195555536015e-08, "loss": 2.2519, "step": 2280 }, { "epoch": 2.224926971762415, "grad_norm": 1.785593606024491, "learning_rate": 5.213677866937508e-08, "loss": 2.2231, "step": 2285 }, { "epoch": 2.2297955209347613, "grad_norm": 1.7400661751479558, "learning_rate": 5.207336035211581e-08, "loss": 2.2306, "step": 2290 }, { "epoch": 2.234664070107108, "grad_norm": 1.7513785103045856, "learning_rate": 5.201165778383047e-08, "loss": 2.184, "step": 2295 }, { "epoch": 2.239532619279455, "grad_norm": 1.8372816676992572, "learning_rate": 5.19516290740677e-08, "loss": 2.1913, "step": 2300 }, { "epoch": 2.2444011684518013, "grad_norm": 1.7411317592151172, "learning_rate": 5.189323324398133e-08, "loss": 2.1749, "step": 2305 }, { "epoch": 2.249269717624148, "grad_norm": 1.7262426588954412, "learning_rate": 5.1836430208924614e-08, "loss": 2.1902, "step": 2310 }, { "epoch": 2.2541382667964944, "grad_norm": 1.7496058178900977, "learning_rate": 5.1781180761330104e-08, "loss": 2.2354, "step": 2315 }, { "epoch": 2.2590068159688412, "grad_norm": 1.7515829381648897, "learning_rate": 5.1727446553871265e-08, "loss": 2.1717, "step": 2320 }, { "epoch": 2.263875365141188, "grad_norm": 1.7101981117837521, "learning_rate": 5.1675190082901985e-08, "loss": 2.1983, "step": 2325 }, { "epoch": 2.2687439143135344, "grad_norm": 1.674731414303109, "learning_rate": 5.1624374672169996e-08, "loss": 2.1964, "step": 2330 }, { "epoch": 2.273612463485881, "grad_norm": 1.8194729895132122, "learning_rate": 5.157496445680062e-08, "loss": 2.208, "step": 2335 }, { "epoch": 2.278481012658228, "grad_norm": 1.7198751813449369, "learning_rate": 5.152692436754697e-08, "loss": 2.2088, "step": 2340 }, { "epoch": 2.2833495618305744, "grad_norm": 1.7052083287832007, "learning_rate": 5.1480220115302935e-08, "loss": 2.2412, "step": 2345 }, { "epoch": 2.288218111002921, "grad_norm": 1.6994269312553734, "learning_rate": 5.143481817587523e-08, "loss": 2.2523, "step": 2350 }, { "epoch": 2.293086660175268, "grad_norm": 1.7821059623881625, "learning_rate": 5.139068577501104e-08, "loss": 2.2306, "step": 2355 }, { "epoch": 2.2979552093476143, "grad_norm": 1.7655828205712845, "learning_rate": 5.134779087367746e-08, "loss": 2.2192, "step": 2360 }, { "epoch": 2.302823758519961, "grad_norm": 1.764331702411005, "learning_rate": 5.130610215358936e-08, "loss": 2.2528, "step": 2365 }, { "epoch": 2.3076923076923075, "grad_norm": 1.814774977446024, "learning_rate": 5.126558900298217e-08, "loss": 2.2311, "step": 2370 }, { "epoch": 2.3125608568646543, "grad_norm": 1.7749398004090056, "learning_rate": 5.122622150262591e-08, "loss": 2.1965, "step": 2375 }, { "epoch": 2.317429406037001, "grad_norm": 1.7732754323290614, "learning_rate": 5.118797041207741e-08, "loss": 2.2307, "step": 2380 }, { "epoch": 2.3222979552093475, "grad_norm": 1.7204661711193003, "learning_rate": 5.1150807156166916e-08, "loss": 2.1813, "step": 2385 }, { "epoch": 2.3271665043816943, "grad_norm": 1.7830208677372077, "learning_rate": 5.111470381171611e-08, "loss": 2.1976, "step": 2390 }, { "epoch": 2.332035053554041, "grad_norm": 1.7100457234343944, "learning_rate": 5.1079633094484e-08, "loss": 2.2421, "step": 2395 }, { "epoch": 2.3369036027263874, "grad_norm": 1.7579822935018907, "learning_rate": 5.104556834633745e-08, "loss": 2.2063, "step": 2400 }, { "epoch": 2.3369036027263874, "eval_loss": 2.4292638301849365, "eval_runtime": 85.8612, "eval_samples_per_second": 85.009, "eval_steps_per_second": 0.676, "step": 2400 }, { "epoch": 2.3417721518987342, "grad_norm": 1.7345007813133775, "learning_rate": 5.101248352264326e-08, "loss": 2.2243, "step": 2405 }, { "epoch": 2.346640701071081, "grad_norm": 1.7248955210954013, "learning_rate": 5.098035317987838e-08, "loss": 2.2096, "step": 2410 }, { "epoch": 2.3515092502434274, "grad_norm": 1.7460382186199548, "learning_rate": 5.0949152463455285e-08, "loss": 2.2034, "step": 2415 }, { "epoch": 2.356377799415774, "grad_norm": 1.7441544396124007, "learning_rate": 5.09188570957593e-08, "loss": 2.1584, "step": 2420 }, { "epoch": 2.3612463485881205, "grad_norm": 1.7037153131860179, "learning_rate": 5.0889443364394804e-08, "loss": 2.222, "step": 2425 }, { "epoch": 2.3661148977604674, "grad_norm": 1.7280060531644463, "learning_rate": 5.0860888110637265e-08, "loss": 2.2324, "step": 2430 }, { "epoch": 2.370983446932814, "grad_norm": 1.7906585557070347, "learning_rate": 5.083316871808814e-08, "loss": 2.2047, "step": 2435 }, { "epoch": 2.3758519961051605, "grad_norm": 1.7845861555659082, "learning_rate": 5.080626310152955e-08, "loss": 2.1902, "step": 2440 }, { "epoch": 2.3807205452775073, "grad_norm": 1.6806913458915649, "learning_rate": 5.078014969597595e-08, "loss": 2.2, "step": 2445 }, { "epoch": 2.385589094449854, "grad_norm": 1.7905906194923593, "learning_rate": 5.075480744591971e-08, "loss": 2.1989, "step": 2450 }, { "epoch": 2.3904576436222005, "grad_norm": 1.7807103616717044, "learning_rate": 5.073021579476786e-08, "loss": 2.2194, "step": 2455 }, { "epoch": 2.3953261927945473, "grad_norm": 1.6957268999562412, "learning_rate": 5.070635467446715e-08, "loss": 2.1982, "step": 2460 }, { "epoch": 2.400194741966894, "grad_norm": 1.775675901114813, "learning_rate": 5.0683204495314504e-08, "loss": 2.213, "step": 2465 }, { "epoch": 2.4050632911392404, "grad_norm": 1.7829403999278695, "learning_rate": 5.06607461359503e-08, "loss": 2.2226, "step": 2470 }, { "epoch": 2.4099318403115872, "grad_norm": 1.766705522512629, "learning_rate": 5.06389609335315e-08, "loss": 2.1896, "step": 2475 }, { "epoch": 2.4148003894839336, "grad_norm": 1.7827496357552255, "learning_rate": 5.0617830674082116e-08, "loss": 2.1947, "step": 2480 }, { "epoch": 2.4196689386562804, "grad_norm": 1.7451999252246684, "learning_rate": 5.059733758301827e-08, "loss": 2.2393, "step": 2485 }, { "epoch": 2.424537487828627, "grad_norm": 1.721359686507254, "learning_rate": 5.057746431584517e-08, "loss": 2.2388, "step": 2490 }, { "epoch": 2.4294060370009736, "grad_norm": 1.7903488742098908, "learning_rate": 5.055819394902345e-08, "loss": 2.1765, "step": 2495 }, { "epoch": 2.4342745861733204, "grad_norm": 1.7278059859253259, "learning_rate": 5.053950997100227e-08, "loss": 2.188, "step": 2500 }, { "epoch": 2.439143135345667, "grad_norm": 1.723905932162126, "learning_rate": 5.052139627341665e-08, "loss": 2.1994, "step": 2505 }, { "epoch": 2.4440116845180135, "grad_norm": 1.8796866769396487, "learning_rate": 5.050383714244649e-08, "loss": 2.2311, "step": 2510 }, { "epoch": 2.4488802336903603, "grad_norm": 1.7459716379486228, "learning_rate": 5.0486817250334816e-08, "loss": 2.2485, "step": 2515 }, { "epoch": 2.453748782862707, "grad_norm": 1.7365026105109531, "learning_rate": 5.047032164706284e-08, "loss": 2.1676, "step": 2520 }, { "epoch": 2.4586173320350535, "grad_norm": 1.7081299835491885, "learning_rate": 5.045433575217931e-08, "loss": 2.1875, "step": 2525 }, { "epoch": 2.4634858812074003, "grad_norm": 1.744755862782243, "learning_rate": 5.043884534678184e-08, "loss": 2.245, "step": 2530 }, { "epoch": 2.4683544303797467, "grad_norm": 1.7422537825667799, "learning_rate": 5.042383656564784e-08, "loss": 2.217, "step": 2535 }, { "epoch": 2.4732229795520935, "grad_norm": 1.734841704470881, "learning_rate": 5.040929588951272e-08, "loss": 2.2173, "step": 2540 }, { "epoch": 2.4780915287244403, "grad_norm": 1.779810459600651, "learning_rate": 5.039521013749303e-08, "loss": 2.247, "step": 2545 }, { "epoch": 2.4829600778967866, "grad_norm": 1.8117164085811932, "learning_rate": 5.0381566459652284e-08, "loss": 2.237, "step": 2550 }, { "epoch": 2.4878286270691334, "grad_norm": 1.7287658099278131, "learning_rate": 5.0368352329707235e-08, "loss": 2.1902, "step": 2555 }, { "epoch": 2.49269717624148, "grad_norm": 1.7649214315832387, "learning_rate": 5.0355555537872345e-08, "loss": 2.2109, "step": 2560 }, { "epoch": 2.4975657254138266, "grad_norm": 1.662642992634986, "learning_rate": 5.0343164183840344e-08, "loss": 2.197, "step": 2565 }, { "epoch": 2.5024342745861734, "grad_norm": 1.7337104838406339, "learning_rate": 5.033116666989654e-08, "loss": 2.2101, "step": 2570 }, { "epoch": 2.50730282375852, "grad_norm": 1.8134178787153215, "learning_rate": 5.031955169416503e-08, "loss": 2.1644, "step": 2575 }, { "epoch": 2.5121713729308666, "grad_norm": 1.780434641418463, "learning_rate": 5.0308308243984355e-08, "loss": 2.2137, "step": 2580 }, { "epoch": 2.5170399221032134, "grad_norm": 1.744873235024722, "learning_rate": 5.0297425589410844e-08, "loss": 2.1802, "step": 2585 }, { "epoch": 2.5219084712755597, "grad_norm": 1.7946463871437135, "learning_rate": 5.0286893276847386e-08, "loss": 2.2194, "step": 2590 }, { "epoch": 2.5267770204479065, "grad_norm": 1.7550387629748228, "learning_rate": 5.0276701122795665e-08, "loss": 2.1656, "step": 2595 }, { "epoch": 2.5316455696202533, "grad_norm": 1.7656891772283707, "learning_rate": 5.02668392077299e-08, "loss": 2.2544, "step": 2600 }, { "epoch": 2.5316455696202533, "eval_loss": 2.429072141647339, "eval_runtime": 85.9551, "eval_samples_per_second": 84.916, "eval_steps_per_second": 0.675, "step": 2600 }, { "epoch": 2.5365141187925997, "grad_norm": 1.8122468666175253, "learning_rate": 5.025729787009003e-08, "loss": 2.2305, "step": 2605 }, { "epoch": 2.5413826679649465, "grad_norm": 1.6973041191639477, "learning_rate": 5.024806770039247e-08, "loss": 2.2292, "step": 2610 }, { "epoch": 2.546251217137293, "grad_norm": 1.729259299213179, "learning_rate": 5.023913953545651e-08, "loss": 2.2614, "step": 2615 }, { "epoch": 2.5511197663096397, "grad_norm": 1.7611375878198783, "learning_rate": 5.023050445274437e-08, "loss": 2.2225, "step": 2620 }, { "epoch": 2.5559883154819865, "grad_norm": 1.7936750412570266, "learning_rate": 5.022215376481317e-08, "loss": 2.2203, "step": 2625 }, { "epoch": 2.5608568646543333, "grad_norm": 1.7697623039425605, "learning_rate": 5.021407901387688e-08, "loss": 2.1863, "step": 2630 }, { "epoch": 2.5657254138266796, "grad_norm": 1.7467899336813517, "learning_rate": 5.02062719664764e-08, "loss": 2.2103, "step": 2635 }, { "epoch": 2.5705939629990264, "grad_norm": 1.7889652968951377, "learning_rate": 5.019872460825613e-08, "loss": 2.2229, "step": 2640 }, { "epoch": 2.575462512171373, "grad_norm": 1.722371905671811, "learning_rate": 5.019142913884503e-08, "loss": 2.1491, "step": 2645 }, { "epoch": 2.5803310613437196, "grad_norm": 1.7502842489092898, "learning_rate": 5.018437796684058e-08, "loss": 2.2225, "step": 2650 }, { "epoch": 2.5851996105160664, "grad_norm": 1.733643511652611, "learning_rate": 5.0177563704893894e-08, "loss": 2.2308, "step": 2655 }, { "epoch": 2.5900681596884128, "grad_norm": 1.7507169934067448, "learning_rate": 5.0170979164894177e-08, "loss": 2.2319, "step": 2660 }, { "epoch": 2.5949367088607596, "grad_norm": 1.7597206213080367, "learning_rate": 5.016461735325101e-08, "loss": 2.2221, "step": 2665 }, { "epoch": 2.599805258033106, "grad_norm": 1.7484941859889187, "learning_rate": 5.0158471466272625e-08, "loss": 2.2408, "step": 2670 }, { "epoch": 2.6046738072054527, "grad_norm": 1.7575142192005702, "learning_rate": 5.0152534885638713e-08, "loss": 2.2199, "step": 2675 }, { "epoch": 2.6095423563777995, "grad_norm": 1.7220313810152217, "learning_rate": 5.014680117396598e-08, "loss": 2.172, "step": 2680 }, { "epoch": 2.6144109055501463, "grad_norm": 1.7246730790979932, "learning_rate": 5.0141264070464985e-08, "loss": 2.2053, "step": 2685 }, { "epoch": 2.6192794547224927, "grad_norm": 1.733539469198201, "learning_rate": 5.013591748668665e-08, "loss": 2.1883, "step": 2690 }, { "epoch": 2.6241480038948395, "grad_norm": 1.708722183137744, "learning_rate": 5.0130755502356856e-08, "loss": 2.2084, "step": 2695 }, { "epoch": 2.629016553067186, "grad_norm": 1.743046661098172, "learning_rate": 5.0125772361297664e-08, "loss": 2.1743, "step": 2700 }, { "epoch": 2.6338851022395326, "grad_norm": 1.8128256440501649, "learning_rate": 5.0120962467433614e-08, "loss": 2.2052, "step": 2705 }, { "epoch": 2.6387536514118795, "grad_norm": 1.7757973503695443, "learning_rate": 5.0116320380881606e-08, "loss": 2.2504, "step": 2710 }, { "epoch": 2.643622200584226, "grad_norm": 1.7774946391866604, "learning_rate": 5.0111840814122985e-08, "loss": 2.1926, "step": 2715 }, { "epoch": 2.6484907497565726, "grad_norm": 1.7302618686374809, "learning_rate": 5.010751862825623e-08, "loss": 2.1849, "step": 2720 }, { "epoch": 2.653359298928919, "grad_norm": 1.7071516593785698, "learning_rate": 5.0103348829328986e-08, "loss": 2.1692, "step": 2725 }, { "epoch": 2.6582278481012658, "grad_norm": 1.721147660276721, "learning_rate": 5.009932656474795e-08, "loss": 2.2232, "step": 2730 }, { "epoch": 2.6630963972736126, "grad_norm": 1.8757024025402924, "learning_rate": 5.0095447119765244e-08, "loss": 2.193, "step": 2735 }, { "epoch": 2.667964946445959, "grad_norm": 1.7734949594595664, "learning_rate": 5.009170591403991e-08, "loss": 2.2176, "step": 2740 }, { "epoch": 2.6728334956183057, "grad_norm": 1.7300923376938375, "learning_rate": 5.0088098498273176e-08, "loss": 2.2405, "step": 2745 }, { "epoch": 2.6777020447906525, "grad_norm": 1.715381643484051, "learning_rate": 5.008462055091624e-08, "loss": 2.1716, "step": 2750 }, { "epoch": 2.682570593962999, "grad_norm": 1.7201490901820933, "learning_rate": 5.00812678749491e-08, "loss": 2.2163, "step": 2755 }, { "epoch": 2.6874391431353457, "grad_norm": 1.776422114245558, "learning_rate": 5.0078036394729376e-08, "loss": 2.2385, "step": 2760 }, { "epoch": 2.6923076923076925, "grad_norm": 1.7434342909806673, "learning_rate": 5.007492215290964e-08, "loss": 2.2148, "step": 2765 }, { "epoch": 2.697176241480039, "grad_norm": 1.7303585866040545, "learning_rate": 5.0071921307422136e-08, "loss": 2.2202, "step": 2770 }, { "epoch": 2.7020447906523857, "grad_norm": 1.7653235238408616, "learning_rate": 5.00690301285296e-08, "loss": 2.2438, "step": 2775 }, { "epoch": 2.706913339824732, "grad_norm": 1.732574656691103, "learning_rate": 5.006624499594101e-08, "loss": 2.1988, "step": 2780 }, { "epoch": 2.711781888997079, "grad_norm": 1.7030324021473955, "learning_rate": 5.0063562395991076e-08, "loss": 2.2623, "step": 2785 }, { "epoch": 2.7166504381694256, "grad_norm": 1.9331342916421579, "learning_rate": 5.006097891888214e-08, "loss": 2.1995, "step": 2790 }, { "epoch": 2.721518987341772, "grad_norm": 1.7201710098478449, "learning_rate": 5.005849125598762e-08, "loss": 2.2411, "step": 2795 }, { "epoch": 2.726387536514119, "grad_norm": 1.6978407566709903, "learning_rate": 5.0056096197215594e-08, "loss": 2.2024, "step": 2800 }, { "epoch": 2.726387536514119, "eval_loss": 2.4288954734802246, "eval_runtime": 85.7694, "eval_samples_per_second": 85.1, "eval_steps_per_second": 0.676, "step": 2800 }, { "epoch": 2.731256085686465, "grad_norm": 1.7625821445707386, "learning_rate": 5.005379062843146e-08, "loss": 2.245, "step": 2805 }, { "epoch": 2.736124634858812, "grad_norm": 1.7304216492048587, "learning_rate": 5.005157152893869e-08, "loss": 2.223, "step": 2810 }, { "epoch": 2.7409931840311588, "grad_norm": 1.755183680113624, "learning_rate": 5.0049435969016435e-08, "loss": 2.2239, "step": 2815 }, { "epoch": 2.7458617332035056, "grad_norm": 1.728272875903093, "learning_rate": 5.004738110751295e-08, "loss": 2.237, "step": 2820 }, { "epoch": 2.750730282375852, "grad_norm": 1.764407536913778, "learning_rate": 5.0045404189493816e-08, "loss": 2.2193, "step": 2825 }, { "epoch": 2.7555988315481987, "grad_norm": 1.7145100854618731, "learning_rate": 5.0043502543943914e-08, "loss": 2.1959, "step": 2830 }, { "epoch": 2.760467380720545, "grad_norm": 1.7322275336580597, "learning_rate": 5.0041673581522057e-08, "loss": 2.1898, "step": 2835 }, { "epoch": 2.765335929892892, "grad_norm": 1.7454769702317845, "learning_rate": 5.003991479236733e-08, "loss": 2.2852, "step": 2840 }, { "epoch": 2.7702044790652387, "grad_norm": 1.8214614189910026, "learning_rate": 5.003822374395615e-08, "loss": 2.2305, "step": 2845 }, { "epoch": 2.775073028237585, "grad_norm": 1.8310587465226418, "learning_rate": 5.0036598079009014e-08, "loss": 2.1533, "step": 2850 }, { "epoch": 2.779941577409932, "grad_norm": 1.773580257235929, "learning_rate": 5.003503551344602e-08, "loss": 2.2247, "step": 2855 }, { "epoch": 2.7848101265822782, "grad_norm": 1.6972419215418109, "learning_rate": 5.003353383439017e-08, "loss": 2.2464, "step": 2860 }, { "epoch": 2.789678675754625, "grad_norm": 1.7541525971250715, "learning_rate": 5.003209089821755e-08, "loss": 2.2041, "step": 2865 }, { "epoch": 2.794547224926972, "grad_norm": 1.7726611140742992, "learning_rate": 5.003070462865347e-08, "loss": 2.2351, "step": 2870 }, { "epoch": 2.7994157740993186, "grad_norm": 1.7257606879168454, "learning_rate": 5.002937301491361e-08, "loss": 2.1964, "step": 2875 }, { "epoch": 2.804284323271665, "grad_norm": 1.7665569916717356, "learning_rate": 5.002809410988936e-08, "loss": 2.2198, "step": 2880 }, { "epoch": 2.809152872444012, "grad_norm": 0.8460817495380386, "learning_rate": 5.00268660283764e-08, "loss": 2.2017, "step": 2885 }, { "epoch": 2.814021421616358, "grad_norm": 1.7781068744183743, "learning_rate": 5.002568694534567e-08, "loss": 2.2113, "step": 2890 }, { "epoch": 2.818889970788705, "grad_norm": 1.70436430018188, "learning_rate": 5.0024555094256e-08, "loss": 2.2296, "step": 2895 }, { "epoch": 2.8237585199610518, "grad_norm": 1.801068375798501, "learning_rate": 5.002346876540729e-08, "loss": 2.2755, "step": 2900 }, { "epoch": 2.828627069133398, "grad_norm": 1.7984300376552464, "learning_rate": 5.0022426304333776e-08, "loss": 2.2464, "step": 2905 }, { "epoch": 2.833495618305745, "grad_norm": 1.7278702590224992, "learning_rate": 5.002142611023626e-08, "loss": 2.2223, "step": 2910 }, { "epoch": 2.8383641674780913, "grad_norm": 1.7544765756817247, "learning_rate": 5.0020466634452695e-08, "loss": 2.2096, "step": 2915 }, { "epoch": 2.843232716650438, "grad_norm": 1.7574601723708598, "learning_rate": 5.001954637896626e-08, "loss": 2.2034, "step": 2920 }, { "epoch": 2.848101265822785, "grad_norm": 1.6850076585078704, "learning_rate": 5.0018663894950185e-08, "loss": 2.1788, "step": 2925 }, { "epoch": 2.8529698149951317, "grad_norm": 1.734136831684721, "learning_rate": 5.001781778134857e-08, "loss": 2.171, "step": 2930 }, { "epoch": 2.857838364167478, "grad_norm": 1.8530752594311186, "learning_rate": 5.0017006683492444e-08, "loss": 2.2323, "step": 2935 }, { "epoch": 2.862706913339825, "grad_norm": 1.7452631968653791, "learning_rate": 5.001622929175032e-08, "loss": 2.2002, "step": 2940 }, { "epoch": 2.867575462512171, "grad_norm": 1.765852324361209, "learning_rate": 5.0015484340212554e-08, "loss": 2.173, "step": 2945 }, { "epoch": 2.872444011684518, "grad_norm": 1.7504024622386836, "learning_rate": 5.0014770605408766e-08, "loss": 2.1818, "step": 2950 }, { "epoch": 2.877312560856865, "grad_norm": 1.709084582717418, "learning_rate": 5.001408690505769e-08, "loss": 2.2248, "step": 2955 }, { "epoch": 2.882181110029211, "grad_norm": 1.726469521262587, "learning_rate": 5.0013432096848647e-08, "loss": 2.1754, "step": 2960 }, { "epoch": 2.887049659201558, "grad_norm": 1.735487847062397, "learning_rate": 5.001280507725414e-08, "loss": 2.1754, "step": 2965 }, { "epoch": 2.8919182083739043, "grad_norm": 1.7901732327186204, "learning_rate": 5.0012204780372755e-08, "loss": 2.2104, "step": 2970 }, { "epoch": 2.896786757546251, "grad_norm": 1.7497305238142942, "learning_rate": 5.001163017680177e-08, "loss": 2.2131, "step": 2975 }, { "epoch": 2.901655306718598, "grad_norm": 1.6912311894555403, "learning_rate": 5.0011080272538895e-08, "loss": 2.1984, "step": 2980 }, { "epoch": 2.9065238558909448, "grad_norm": 1.7550540502025136, "learning_rate": 5.0010657486604285e-08, "loss": 2.2325, "step": 2985 }, { "epoch": 2.911392405063291, "grad_norm": 1.7830928527818988, "learning_rate": 5.001014964512753e-08, "loss": 2.2295, "step": 2990 }, { "epoch": 2.916260954235638, "grad_norm": 1.7967938470353484, "learning_rate": 5.000966389908247e-08, "loss": 2.2238, "step": 2995 }, { "epoch": 2.9211295034079843, "grad_norm": 1.73275794534534, "learning_rate": 5.0009199379466084e-08, "loss": 2.2074, "step": 3000 }, { "epoch": 2.9211295034079843, "eval_loss": 2.4288480281829834, "eval_runtime": 85.4779, "eval_samples_per_second": 85.391, "eval_steps_per_second": 0.679, "step": 3000 }, { "epoch": 2.925998052580331, "grad_norm": 1.798848636728742, "learning_rate": 5.000875524778925e-08, "loss": 2.1889, "step": 3005 }, { "epoch": 2.930866601752678, "grad_norm": 1.7669718423903236, "learning_rate": 5.0008330695135104e-08, "loss": 2.2008, "step": 3010 }, { "epoch": 2.9357351509250242, "grad_norm": 1.7532733346720555, "learning_rate": 5.000792494124235e-08, "loss": 2.193, "step": 3015 }, { "epoch": 2.940603700097371, "grad_norm": 1.7095158480632124, "learning_rate": 5.00075372336131e-08, "loss": 2.2389, "step": 3020 }, { "epoch": 2.9454722492697174, "grad_norm": 1.7236728786130506, "learning_rate": 5.0007166846644555e-08, "loss": 2.2444, "step": 3025 }, { "epoch": 2.950340798442064, "grad_norm": 1.7363740627339725, "learning_rate": 5.000681308078417e-08, "loss": 2.22, "step": 3030 }, { "epoch": 2.955209347614411, "grad_norm": 1.728808934644446, "learning_rate": 5.000647526170754e-08, "loss": 2.2163, "step": 3035 }, { "epoch": 2.960077896786758, "grad_norm": 1.7449413432648693, "learning_rate": 5.000615273951875e-08, "loss": 2.2779, "step": 3040 }, { "epoch": 2.964946445959104, "grad_norm": 1.753647141186398, "learning_rate": 5.000584488797249e-08, "loss": 2.1997, "step": 3045 }, { "epoch": 2.969814995131451, "grad_norm": 1.729320335247063, "learning_rate": 5.000555110371747e-08, "loss": 2.2329, "step": 3050 }, { "epoch": 2.9746835443037973, "grad_norm": 1.7435846324509203, "learning_rate": 5.000527080556074e-08, "loss": 2.2455, "step": 3055 }, { "epoch": 2.979552093476144, "grad_norm": 1.7544585151474008, "learning_rate": 5.000500343375234e-08, "loss": 2.2232, "step": 3060 }, { "epoch": 2.984420642648491, "grad_norm": 1.8297856934072276, "learning_rate": 5.00047484492898e-08, "loss": 2.2025, "step": 3065 }, { "epoch": 2.9892891918208373, "grad_norm": 1.8331124463008972, "learning_rate": 5.000450533324217e-08, "loss": 2.2346, "step": 3070 }, { "epoch": 2.994157740993184, "grad_norm": 1.77523332941347, "learning_rate": 5.0004273586092874e-08, "loss": 2.2476, "step": 3075 }, { "epoch": 2.9990262901655305, "grad_norm": 1.719560891287608, "learning_rate": 5.0004052727101194e-08, "loss": 2.2219, "step": 3080 }, { "epoch": 3.0038948393378773, "grad_norm": 1.7451899184767314, "learning_rate": 5.000384229368183e-08, "loss": 2.267, "step": 3085 }, { "epoch": 3.008763388510224, "grad_norm": 1.7132088078853227, "learning_rate": 5.0003641840802084e-08, "loss": 2.232, "step": 3090 }, { "epoch": 3.0136319376825704, "grad_norm": 1.7074710408698504, "learning_rate": 5.000345094039628e-08, "loss": 2.2226, "step": 3095 }, { "epoch": 3.0185004868549172, "grad_norm": 1.7116453045970612, "learning_rate": 5.000326918079705e-08, "loss": 2.2132, "step": 3100 }, { "epoch": 3.023369036027264, "grad_norm": 1.7478082412483065, "learning_rate": 5.000309616618301e-08, "loss": 2.2595, "step": 3105 }, { "epoch": 3.0282375851996104, "grad_norm": 1.7390659451407764, "learning_rate": 5.0002931516042464e-08, "loss": 2.2532, "step": 3110 }, { "epoch": 3.033106134371957, "grad_norm": 1.6629867778767704, "learning_rate": 5.0002774864652723e-08, "loss": 2.2157, "step": 3115 }, { "epoch": 3.037974683544304, "grad_norm": 1.7430106410373127, "learning_rate": 5.000262586057472e-08, "loss": 2.2654, "step": 3120 }, { "epoch": 3.0428432327166504, "grad_norm": 1.7699263854675031, "learning_rate": 5.0002484166162455e-08, "loss": 2.1838, "step": 3125 }, { "epoch": 3.047711781888997, "grad_norm": 1.7970379405255692, "learning_rate": 5.0002349457087e-08, "loss": 2.1844, "step": 3130 }, { "epoch": 3.0525803310613435, "grad_norm": 1.740063554373285, "learning_rate": 5.00022214218746e-08, "loss": 2.1975, "step": 3135 }, { "epoch": 3.0574488802336903, "grad_norm": 1.7669764533755763, "learning_rate": 5.000209976145863e-08, "loss": 2.173, "step": 3140 }, { "epoch": 3.062317429406037, "grad_norm": 1.7570288798515004, "learning_rate": 5.000198418874495e-08, "loss": 2.202, "step": 3145 }, { "epoch": 3.0671859785783835, "grad_norm": 1.7640313096994278, "learning_rate": 5.000187442819041e-08, "loss": 2.2743, "step": 3150 }, { "epoch": 3.0720545277507303, "grad_norm": 1.776697414652845, "learning_rate": 5.000177021539416e-08, "loss": 2.1953, "step": 3155 }, { "epoch": 3.076923076923077, "grad_norm": 1.7469177371601763, "learning_rate": 5.000167129670136e-08, "loss": 2.2643, "step": 3160 }, { "epoch": 3.0817916260954235, "grad_norm": 1.7981065248029078, "learning_rate": 5.000157742881906e-08, "loss": 2.2192, "step": 3165 }, { "epoch": 3.0866601752677703, "grad_norm": 1.8048538458330963, "learning_rate": 5.000148837844394e-08, "loss": 2.2188, "step": 3170 }, { "epoch": 3.091528724440117, "grad_norm": 1.77587620793656, "learning_rate": 5.000140392190154e-08, "loss": 2.205, "step": 3175 }, { "epoch": 3.0963972736124634, "grad_norm": 1.7277030937813718, "learning_rate": 5.000132384479669e-08, "loss": 2.2418, "step": 3180 }, { "epoch": 3.1012658227848102, "grad_norm": 1.7715455250981111, "learning_rate": 5.0001247941674944e-08, "loss": 2.195, "step": 3185 }, { "epoch": 3.1061343719571566, "grad_norm": 1.7418270265834659, "learning_rate": 5.00011760156946e-08, "loss": 2.2188, "step": 3190 }, { "epoch": 3.1110029211295034, "grad_norm": 1.7626475968666995, "learning_rate": 5.000110787830913e-08, "loss": 2.1969, "step": 3195 }, { "epoch": 3.11587147030185, "grad_norm": 1.7554268739931094, "learning_rate": 5.000104334895965e-08, "loss": 2.2268, "step": 3200 }, { "epoch": 3.11587147030185, "eval_loss": 2.4296844005584717, "eval_runtime": 85.8007, "eval_samples_per_second": 85.069, "eval_steps_per_second": 0.676, "step": 3200 }, { "epoch": 3.1207400194741965, "grad_norm": 1.768111175754245, "learning_rate": 5.000098225477726e-08, "loss": 2.1925, "step": 3205 }, { "epoch": 3.1256085686465434, "grad_norm": 1.731673904293537, "learning_rate": 5.000092443029491e-08, "loss": 2.2291, "step": 3210 }, { "epoch": 3.13047711781889, "grad_norm": 1.7474269730832317, "learning_rate": 5.0000869717168615e-08, "loss": 2.2431, "step": 3215 }, { "epoch": 3.1353456669912365, "grad_norm": 1.7068082256216377, "learning_rate": 5.000081796390766e-08, "loss": 2.2197, "step": 3220 }, { "epoch": 3.1402142161635833, "grad_norm": 1.791689784795618, "learning_rate": 5.000076902561367e-08, "loss": 2.1983, "step": 3225 }, { "epoch": 3.14508276533593, "grad_norm": 1.7624390988374115, "learning_rate": 5.000072276372817e-08, "loss": 2.2693, "step": 3230 }, { "epoch": 3.1499513145082765, "grad_norm": 1.8338687668629183, "learning_rate": 5.0000679045788575e-08, "loss": 2.1917, "step": 3235 }, { "epoch": 3.1548198636806233, "grad_norm": 1.709244589398017, "learning_rate": 5.000063774519218e-08, "loss": 2.1911, "step": 3240 }, { "epoch": 3.1596884128529696, "grad_norm": 1.7192540273943981, "learning_rate": 5.0000598740968074e-08, "loss": 2.1583, "step": 3245 }, { "epoch": 3.1645569620253164, "grad_norm": 1.728010839757635, "learning_rate": 5.000056191755672e-08, "loss": 2.1972, "step": 3250 }, { "epoch": 3.1694255111976632, "grad_norm": 1.788796272197145, "learning_rate": 5.0000527164596915e-08, "loss": 2.2011, "step": 3255 }, { "epoch": 3.1742940603700096, "grad_norm": 1.7635432660617776, "learning_rate": 5.000049437672004e-08, "loss": 2.2152, "step": 3260 }, { "epoch": 3.1791626095423564, "grad_norm": 1.7957521394770797, "learning_rate": 5.000046345335129e-08, "loss": 2.2182, "step": 3265 }, { "epoch": 3.184031158714703, "grad_norm": 1.8009009243137868, "learning_rate": 5.000043429851777e-08, "loss": 2.2061, "step": 3270 }, { "epoch": 3.1888997078870496, "grad_norm": 1.7948019721772699, "learning_rate": 5.0000406820663126e-08, "loss": 2.1857, "step": 3275 }, { "epoch": 3.1937682570593964, "grad_norm": 1.8100815667189885, "learning_rate": 5.0000380932468733e-08, "loss": 2.1777, "step": 3280 }, { "epoch": 3.1986368062317427, "grad_norm": 1.7827984808365251, "learning_rate": 5.000035655068104e-08, "loss": 2.2366, "step": 3285 }, { "epoch": 3.2035053554040895, "grad_norm": 1.748286864117926, "learning_rate": 5.0000333595944974e-08, "loss": 2.207, "step": 3290 }, { "epoch": 3.2083739045764363, "grad_norm": 1.703650849390858, "learning_rate": 5.0000311992643325e-08, "loss": 2.1896, "step": 3295 }, { "epoch": 3.2132424537487827, "grad_norm": 1.7698280644275959, "learning_rate": 5.0000291668741705e-08, "loss": 2.1639, "step": 3300 }, { "epoch": 3.2181110029211295, "grad_norm": 1.756366599744074, "learning_rate": 5.000027255563917e-08, "loss": 2.2012, "step": 3305 }, { "epoch": 3.2229795520934763, "grad_norm": 1.759461002068701, "learning_rate": 5.000025458802413e-08, "loss": 2.2341, "step": 3310 }, { "epoch": 3.2278481012658227, "grad_norm": 1.7494660148125987, "learning_rate": 5.000023770373551e-08, "loss": 2.207, "step": 3315 }, { "epoch": 3.2327166504381695, "grad_norm": 1.7785100579340334, "learning_rate": 5.000022184362899e-08, "loss": 2.1999, "step": 3320 }, { "epoch": 3.2375851996105163, "grad_norm": 1.7645954769969932, "learning_rate": 5.0000206951448066e-08, "loss": 2.2683, "step": 3325 }, { "epoch": 3.2424537487828626, "grad_norm": 1.7745088927616617, "learning_rate": 5.000019297369995e-08, "loss": 2.2251, "step": 3330 }, { "epoch": 3.2473222979552094, "grad_norm": 1.8131533062966378, "learning_rate": 5.0000179859535986e-08, "loss": 2.2074, "step": 3335 }, { "epoch": 3.252190847127556, "grad_norm": 1.744211372153281, "learning_rate": 5.000016756063664e-08, "loss": 2.172, "step": 3340 }, { "epoch": 3.2570593962999026, "grad_norm": 1.7741244031891719, "learning_rate": 5.000015603110066e-08, "loss": 2.2425, "step": 3345 }, { "epoch": 3.2619279454722494, "grad_norm": 1.7486152934920032, "learning_rate": 5.0000145227338575e-08, "loss": 2.2143, "step": 3350 }, { "epoch": 3.2667964946445958, "grad_norm": 1.757112828413238, "learning_rate": 5.000013510797011e-08, "loss": 2.1976, "step": 3355 }, { "epoch": 3.2716650438169426, "grad_norm": 1.7240716262707467, "learning_rate": 5.000012563372563e-08, "loss": 2.2618, "step": 3360 }, { "epoch": 3.2765335929892894, "grad_norm": 1.7213155493984982, "learning_rate": 5.0000116767351296e-08, "loss": 2.2115, "step": 3365 }, { "epoch": 3.2814021421616357, "grad_norm": 1.8311141582937929, "learning_rate": 5.000010847351797e-08, "loss": 2.1974, "step": 3370 }, { "epoch": 3.2862706913339825, "grad_norm": 1.7864630099921617, "learning_rate": 5.000010071873363e-08, "loss": 2.238, "step": 3375 }, { "epoch": 3.291139240506329, "grad_norm": 1.7883459145585372, "learning_rate": 5.00000934712592e-08, "loss": 2.2031, "step": 3380 }, { "epoch": 3.2960077896786757, "grad_norm": 1.699599322514979, "learning_rate": 5.000008670102778e-08, "loss": 2.2327, "step": 3385 }, { "epoch": 3.3008763388510225, "grad_norm": 1.7788870047828989, "learning_rate": 5.0000080379567014e-08, "loss": 2.1472, "step": 3390 }, { "epoch": 3.305744888023369, "grad_norm": 1.7365094304977247, "learning_rate": 5.000007447992463e-08, "loss": 2.2068, "step": 3395 }, { "epoch": 3.3106134371957157, "grad_norm": 1.8492451880826386, "learning_rate": 5.0000068976596906e-08, "loss": 2.1556, "step": 3400 }, { "epoch": 3.3106134371957157, "eval_loss": 2.429415464401245, "eval_runtime": 85.0239, "eval_samples_per_second": 85.846, "eval_steps_per_second": 0.682, "step": 3400 }, { "epoch": 3.3154819863680625, "grad_norm": 1.769012775563444, "learning_rate": 5.0000063845460134e-08, "loss": 2.1791, "step": 3405 }, { "epoch": 3.320350535540409, "grad_norm": 1.7426924775202128, "learning_rate": 5.000005999315767e-08, "loss": 2.2323, "step": 3410 }, { "epoch": 3.3252190847127556, "grad_norm": 1.764543573084179, "learning_rate": 5.000005547532745e-08, "loss": 2.1965, "step": 3415 }, { "epoch": 3.3300876338851024, "grad_norm": 1.8373323305401401, "learning_rate": 5.0000051268933666e-08, "loss": 2.2295, "step": 3420 }, { "epoch": 3.334956183057449, "grad_norm": 1.713591544576181, "learning_rate": 5.00000473545193e-08, "loss": 2.1847, "step": 3425 }, { "epoch": 3.3398247322297956, "grad_norm": 1.7559454568881032, "learning_rate": 5.000004371371542e-08, "loss": 2.2148, "step": 3430 }, { "epoch": 3.344693281402142, "grad_norm": 1.7485006911314465, "learning_rate": 5.000004032918764e-08, "loss": 2.2065, "step": 3435 }, { "epoch": 3.3495618305744888, "grad_norm": 1.7533601433403279, "learning_rate": 5.00000371845847e-08, "loss": 2.1955, "step": 3440 }, { "epoch": 3.3544303797468356, "grad_norm": 1.7543658707594614, "learning_rate": 5.000003426448936e-08, "loss": 2.2087, "step": 3445 }, { "epoch": 3.359298928919182, "grad_norm": 1.7904175039614314, "learning_rate": 5.0000031554371375e-08, "loss": 2.272, "step": 3450 }, { "epoch": 3.3641674780915287, "grad_norm": 1.7502721668385577, "learning_rate": 5.000002904054251e-08, "loss": 2.2101, "step": 3455 }, { "epoch": 3.3690360272638755, "grad_norm": 1.706493138085354, "learning_rate": 5.000002671011354e-08, "loss": 2.1955, "step": 3460 }, { "epoch": 3.373904576436222, "grad_norm": 1.7655034010253434, "learning_rate": 5.0000024550953135e-08, "loss": 2.2103, "step": 3465 }, { "epoch": 3.3787731256085687, "grad_norm": 1.7580566741004937, "learning_rate": 5.0000022551648575e-08, "loss": 2.2243, "step": 3470 }, { "epoch": 3.3836416747809155, "grad_norm": 1.7951986660353298, "learning_rate": 5.000002070146821e-08, "loss": 2.233, "step": 3475 }, { "epoch": 3.388510223953262, "grad_norm": 1.7464673450399157, "learning_rate": 5.000001899032566e-08, "loss": 2.1938, "step": 3480 }, { "epoch": 3.3933787731256086, "grad_norm": 1.7506731049190618, "learning_rate": 5.0000017408745575e-08, "loss": 2.2132, "step": 3485 }, { "epoch": 3.398247322297955, "grad_norm": 1.7158313234275915, "learning_rate": 5.0000015947831005e-08, "loss": 2.2255, "step": 3490 }, { "epoch": 3.403115871470302, "grad_norm": 1.7741448257648114, "learning_rate": 5.000001459923229e-08, "loss": 2.2561, "step": 3495 }, { "epoch": 3.4079844206426486, "grad_norm": 1.7680229196146826, "learning_rate": 5.0000013355117324e-08, "loss": 2.2174, "step": 3500 }, { "epoch": 3.412852969814995, "grad_norm": 1.7566868775329945, "learning_rate": 5.000001220814333e-08, "loss": 2.2049, "step": 3505 }, { "epoch": 3.4177215189873418, "grad_norm": 1.7231310116084255, "learning_rate": 5.0000011151429815e-08, "loss": 2.243, "step": 3510 }, { "epoch": 3.4225900681596886, "grad_norm": 1.7525419468509682, "learning_rate": 5.0000010178532987e-08, "loss": 2.2271, "step": 3515 }, { "epoch": 3.427458617332035, "grad_norm": 1.7743431604185957, "learning_rate": 5.000000928342124e-08, "loss": 2.1975, "step": 3520 }, { "epoch": 3.4323271665043817, "grad_norm": 1.738567333307858, "learning_rate": 5.000000846045193e-08, "loss": 2.1717, "step": 3525 }, { "epoch": 3.4371957156767285, "grad_norm": 1.7930958206448022, "learning_rate": 5.000000770434924e-08, "loss": 2.1939, "step": 3530 }, { "epoch": 3.442064264849075, "grad_norm": 1.7498436482409632, "learning_rate": 5.0000007010183126e-08, "loss": 2.213, "step": 3535 }, { "epoch": 3.4469328140214217, "grad_norm": 1.7055196904583134, "learning_rate": 5.0000006373349365e-08, "loss": 2.2058, "step": 3540 }, { "epoch": 3.451801363193768, "grad_norm": 1.7510250220237107, "learning_rate": 5.00000057895505e-08, "loss": 2.2337, "step": 3545 }, { "epoch": 3.456669912366115, "grad_norm": 1.713344136042387, "learning_rate": 5.0000005254777845e-08, "loss": 2.1989, "step": 3550 }, { "epoch": 3.4615384615384617, "grad_norm": 1.7371767245978642, "learning_rate": 5.000000476529434e-08, "loss": 2.233, "step": 3555 }, { "epoch": 3.466407010710808, "grad_norm": 1.7636410740956914, "learning_rate": 5.000000431761833e-08, "loss": 2.1615, "step": 3560 }, { "epoch": 3.471275559883155, "grad_norm": 1.7144161091063708, "learning_rate": 5.000000390850812e-08, "loss": 2.2205, "step": 3565 }, { "epoch": 3.4761441090555016, "grad_norm": 1.8042049963234217, "learning_rate": 5.0000003534947425e-08, "loss": 2.178, "step": 3570 }, { "epoch": 3.481012658227848, "grad_norm": 1.7853951456456785, "learning_rate": 5.00000031941315e-08, "loss": 2.2185, "step": 3575 }, { "epoch": 3.485881207400195, "grad_norm": 1.6707790882813032, "learning_rate": 5.000000288345406e-08, "loss": 2.1868, "step": 3580 }, { "epoch": 3.4907497565725416, "grad_norm": 1.7236640638911755, "learning_rate": 5.000000260049486e-08, "loss": 2.2153, "step": 3585 }, { "epoch": 3.495618305744888, "grad_norm": 1.7103907676750174, "learning_rate": 5.0000002343008e-08, "loss": 2.2806, "step": 3590 }, { "epoch": 3.5004868549172348, "grad_norm": 1.7458900493895921, "learning_rate": 5.000000210891081e-08, "loss": 2.1832, "step": 3595 }, { "epoch": 3.505355404089581, "grad_norm": 1.8597626382974297, "learning_rate": 5.00000018962734e-08, "loss": 2.1953, "step": 3600 }, { "epoch": 3.505355404089581, "eval_loss": 2.4295620918273926, "eval_runtime": 85.4888, "eval_samples_per_second": 85.38, "eval_steps_per_second": 0.678, "step": 3600 }, { "epoch": 3.510223953261928, "grad_norm": 1.7712796854715114, "learning_rate": 5.000000170330872e-08, "loss": 2.1841, "step": 3605 }, { "epoch": 3.5150925024342747, "grad_norm": 1.7646034875776253, "learning_rate": 5.000000152836327e-08, "loss": 2.2182, "step": 3610 }, { "epoch": 3.519961051606621, "grad_norm": 1.7404348200627402, "learning_rate": 5.000000136990825e-08, "loss": 2.2245, "step": 3615 }, { "epoch": 3.524829600778968, "grad_norm": 1.7098638272240443, "learning_rate": 5.000000122653126e-08, "loss": 2.2094, "step": 3620 }, { "epoch": 3.5296981499513143, "grad_norm": 1.7920393050904846, "learning_rate": 5.000000109692848e-08, "loss": 2.2366, "step": 3625 }, { "epoch": 3.534566699123661, "grad_norm": 1.746241513422538, "learning_rate": 5.0000000979897296e-08, "loss": 2.1925, "step": 3630 }, { "epoch": 3.539435248296008, "grad_norm": 1.8183591426270245, "learning_rate": 5.000000087432932e-08, "loss": 2.2159, "step": 3635 }, { "epoch": 3.5443037974683547, "grad_norm": 1.821208130477108, "learning_rate": 5.0000000779203936e-08, "loss": 2.2038, "step": 3640 }, { "epoch": 3.549172346640701, "grad_norm": 1.7255936567126264, "learning_rate": 5.000000069358209e-08, "loss": 2.1901, "step": 3645 }, { "epoch": 3.554040895813048, "grad_norm": 1.855947848851622, "learning_rate": 5.0000000616600557e-08, "loss": 2.2198, "step": 3650 }, { "epoch": 3.558909444985394, "grad_norm": 1.8303786214402666, "learning_rate": 5.000000054746654e-08, "loss": 2.1946, "step": 3655 }, { "epoch": 3.563777994157741, "grad_norm": 1.7392508415046033, "learning_rate": 5.000000048545256e-08, "loss": 2.1778, "step": 3660 }, { "epoch": 3.568646543330088, "grad_norm": 1.8213274923689418, "learning_rate": 5.0000000429891724e-08, "loss": 2.1742, "step": 3665 }, { "epoch": 3.573515092502434, "grad_norm": 1.8017867977855504, "learning_rate": 5.0000000380173246e-08, "loss": 2.1915, "step": 3670 }, { "epoch": 3.578383641674781, "grad_norm": 1.7382969587337458, "learning_rate": 5.000000033573829e-08, "loss": 2.1693, "step": 3675 }, { "epoch": 3.5832521908471273, "grad_norm": 1.6944625431597526, "learning_rate": 5.0000000296076037e-08, "loss": 2.1787, "step": 3680 }, { "epoch": 3.588120740019474, "grad_norm": 1.726132620813893, "learning_rate": 5.0000000260720054e-08, "loss": 2.2046, "step": 3685 }, { "epoch": 3.592989289191821, "grad_norm": 1.715704002261475, "learning_rate": 5.0000000229244886e-08, "loss": 2.2242, "step": 3690 }, { "epoch": 3.5978578383641677, "grad_norm": 1.7355446090598186, "learning_rate": 5.0000000201262854e-08, "loss": 2.2162, "step": 3695 }, { "epoch": 3.602726387536514, "grad_norm": 1.744140516288637, "learning_rate": 5.0000000176421085e-08, "loss": 2.1987, "step": 3700 }, { "epoch": 3.607594936708861, "grad_norm": 1.801416253511192, "learning_rate": 5.000000015439878e-08, "loss": 2.2142, "step": 3705 }, { "epoch": 3.6124634858812072, "grad_norm": 1.72380914141294, "learning_rate": 5.000000013490458e-08, "loss": 2.1938, "step": 3710 }, { "epoch": 3.617332035053554, "grad_norm": 1.6829634226373547, "learning_rate": 5.000000011767422e-08, "loss": 2.2291, "step": 3715 }, { "epoch": 3.622200584225901, "grad_norm": 1.71665602485241, "learning_rate": 5.0000000102468287e-08, "loss": 2.2024, "step": 3720 }, { "epoch": 3.627069133398247, "grad_norm": 1.7001891231981483, "learning_rate": 5.000000008907013e-08, "loss": 2.2397, "step": 3725 }, { "epoch": 3.631937682570594, "grad_norm": 1.7560852361384687, "learning_rate": 5.000000007728397e-08, "loss": 2.2149, "step": 3730 }, { "epoch": 3.6368062317429404, "grad_norm": 1.8239331341308023, "learning_rate": 5.0000000066933104e-08, "loss": 2.2073, "step": 3735 }, { "epoch": 3.641674780915287, "grad_norm": 1.7435406529125186, "learning_rate": 5.000000005785826e-08, "loss": 2.202, "step": 3740 }, { "epoch": 3.646543330087634, "grad_norm": 1.769830204490899, "learning_rate": 5.0000000049916066e-08, "loss": 2.1874, "step": 3745 }, { "epoch": 3.651411879259981, "grad_norm": 1.7365544971501097, "learning_rate": 5.0000000042977675e-08, "loss": 2.2312, "step": 3750 }, { "epoch": 3.656280428432327, "grad_norm": 1.754918261553824, "learning_rate": 5.000000003692742e-08, "loss": 2.233, "step": 3755 }, { "epoch": 3.661148977604674, "grad_norm": 1.7251476744734402, "learning_rate": 5.000000003166163e-08, "loss": 2.1785, "step": 3760 }, { "epoch": 3.6660175267770203, "grad_norm": 1.6885151956951188, "learning_rate": 5.000000002708756e-08, "loss": 2.213, "step": 3765 }, { "epoch": 3.670886075949367, "grad_norm": 1.8104164593434913, "learning_rate": 5.000000002312232e-08, "loss": 2.2031, "step": 3770 }, { "epoch": 3.675754625121714, "grad_norm": 1.6964800041666812, "learning_rate": 5.0000000019691985e-08, "loss": 2.2106, "step": 3775 }, { "epoch": 3.6806231742940603, "grad_norm": 1.7770021912734775, "learning_rate": 5.0000000016730705e-08, "loss": 2.1958, "step": 3780 }, { "epoch": 3.685491723466407, "grad_norm": 1.7854463027276486, "learning_rate": 5.000000001417995e-08, "loss": 2.1748, "step": 3785 }, { "epoch": 3.6903602726387534, "grad_norm": 1.7314209770363447, "learning_rate": 5.000000001198777e-08, "loss": 2.2285, "step": 3790 }, { "epoch": 3.6952288218111002, "grad_norm": 1.7315368458494336, "learning_rate": 5.000000001010814e-08, "loss": 2.2168, "step": 3795 }, { "epoch": 3.700097370983447, "grad_norm": 1.7953711064417144, "learning_rate": 5.0000000008500364e-08, "loss": 2.2002, "step": 3800 }, { "epoch": 3.700097370983447, "eval_loss": 2.4294352531433105, "eval_runtime": 85.6309, "eval_samples_per_second": 85.238, "eval_steps_per_second": 0.677, "step": 3800 }, { "epoch": 3.704965920155794, "grad_norm": 1.792810318597494, "learning_rate": 5.000000000712853e-08, "loss": 2.2033, "step": 3805 }, { "epoch": 3.70983446932814, "grad_norm": 1.7519531928443899, "learning_rate": 5.0000000005961005e-08, "loss": 2.2504, "step": 3810 }, { "epoch": 3.714703018500487, "grad_norm": 1.8276341931990958, "learning_rate": 5.0000000004969983e-08, "loss": 2.1868, "step": 3815 }, { "epoch": 3.7195715676728334, "grad_norm": 1.7578731153653255, "learning_rate": 5.0000000004131077e-08, "loss": 2.2114, "step": 3820 }, { "epoch": 3.72444011684518, "grad_norm": 1.7261756289309933, "learning_rate": 5.000000000342294e-08, "loss": 2.1714, "step": 3825 }, { "epoch": 3.729308666017527, "grad_norm": 1.7847142724316984, "learning_rate": 5.0000000002826924e-08, "loss": 2.1744, "step": 3830 }, { "epoch": 3.7341772151898733, "grad_norm": 1.7506323765963263, "learning_rate": 5.0000000002326804e-08, "loss": 2.2474, "step": 3835 }, { "epoch": 3.73904576436222, "grad_norm": 1.7726668919984896, "learning_rate": 5.000000000190845e-08, "loss": 2.1832, "step": 3840 }, { "epoch": 3.7439143135345665, "grad_norm": 1.791217388707096, "learning_rate": 5.000000000155963e-08, "loss": 2.2552, "step": 3845 }, { "epoch": 3.7487828627069133, "grad_norm": 1.7260937822564961, "learning_rate": 5.000000000126977e-08, "loss": 2.2135, "step": 3850 }, { "epoch": 3.75365141187926, "grad_norm": 1.8761923429228229, "learning_rate": 5.000000000102974e-08, "loss": 2.2143, "step": 3855 }, { "epoch": 3.7585199610516065, "grad_norm": 1.8000916225190866, "learning_rate": 5.000000000083169e-08, "loss": 2.1963, "step": 3860 }, { "epoch": 3.7633885102239533, "grad_norm": 1.8139182234780402, "learning_rate": 5.0000000000668896e-08, "loss": 2.2333, "step": 3865 }, { "epoch": 3.7682570593963, "grad_norm": 1.779253821129295, "learning_rate": 5.000000000053561e-08, "loss": 2.2217, "step": 3870 }, { "epoch": 3.7731256085686464, "grad_norm": 1.714027215371266, "learning_rate": 5.0000000000426916e-08, "loss": 2.2047, "step": 3875 }, { "epoch": 3.7779941577409932, "grad_norm": 1.803155267019016, "learning_rate": 5.000000000033866e-08, "loss": 2.2063, "step": 3880 }, { "epoch": 3.78286270691334, "grad_norm": 1.7914082013975081, "learning_rate": 5.0000000000267316e-08, "loss": 2.2199, "step": 3885 }, { "epoch": 3.7877312560856864, "grad_norm": 1.7354663482603165, "learning_rate": 5.00000000002099e-08, "loss": 2.2118, "step": 3890 }, { "epoch": 3.792599805258033, "grad_norm": 1.9003792042576655, "learning_rate": 5.000000000016393e-08, "loss": 2.1497, "step": 3895 }, { "epoch": 3.7974683544303796, "grad_norm": 1.76200601661607, "learning_rate": 5.0000000000127304e-08, "loss": 2.1832, "step": 3900 }, { "epoch": 3.8023369036027264, "grad_norm": 1.831527213740064, "learning_rate": 5.0000000000098273e-08, "loss": 2.2266, "step": 3905 }, { "epoch": 3.807205452775073, "grad_norm": 1.796422966992554, "learning_rate": 5.000000000007539e-08, "loss": 2.2171, "step": 3910 }, { "epoch": 3.8120740019474195, "grad_norm": 1.8025926494413154, "learning_rate": 5.000000000005746e-08, "loss": 2.2325, "step": 3915 }, { "epoch": 3.8169425511197663, "grad_norm": 1.793793977270209, "learning_rate": 5.000000000004349e-08, "loss": 2.2545, "step": 3920 }, { "epoch": 3.8218111002921127, "grad_norm": 1.7928832433425952, "learning_rate": 5.000000000003268e-08, "loss": 2.182, "step": 3925 }, { "epoch": 3.8266796494644595, "grad_norm": 1.7947342233214896, "learning_rate": 5.000000000002437e-08, "loss": 2.1897, "step": 3930 }, { "epoch": 3.8315481986368063, "grad_norm": 1.7630875066863383, "learning_rate": 5.000000000001803e-08, "loss": 2.1764, "step": 3935 }, { "epoch": 3.836416747809153, "grad_norm": 1.7899032248006117, "learning_rate": 5.0000000000013226e-08, "loss": 2.1868, "step": 3940 }, { "epoch": 3.8412852969814995, "grad_norm": 1.7615217853308245, "learning_rate": 5.000000000000961e-08, "loss": 2.1874, "step": 3945 }, { "epoch": 3.8461538461538463, "grad_norm": 1.8191785253367547, "learning_rate": 5.000000000000692e-08, "loss": 2.2096, "step": 3950 }, { "epoch": 3.8510223953261926, "grad_norm": 1.7142116315629525, "learning_rate": 5.0000000000004934e-08, "loss": 2.1825, "step": 3955 }, { "epoch": 3.8558909444985394, "grad_norm": 1.8193971497080461, "learning_rate": 5.000000000000348e-08, "loss": 2.1726, "step": 3960 }, { "epoch": 3.8607594936708862, "grad_norm": 1.7556627121596253, "learning_rate": 5.0000000000002426e-08, "loss": 2.2139, "step": 3965 }, { "epoch": 3.8656280428432326, "grad_norm": 1.744589516587126, "learning_rate": 5.000000000000167e-08, "loss": 2.2201, "step": 3970 }, { "epoch": 3.8704965920155794, "grad_norm": 1.714747302449748, "learning_rate": 5.0000000000001136e-08, "loss": 2.2009, "step": 3975 }, { "epoch": 3.8753651411879257, "grad_norm": 1.768710509140083, "learning_rate": 5.000000000000076e-08, "loss": 2.2031, "step": 3980 }, { "epoch": 3.8802336903602725, "grad_norm": 1.7322811102830442, "learning_rate": 5.00000000000005e-08, "loss": 2.2165, "step": 3985 }, { "epoch": 3.8851022395326194, "grad_norm": 1.8064547108979563, "learning_rate": 5.000000000000032e-08, "loss": 2.251, "step": 3990 }, { "epoch": 3.889970788704966, "grad_norm": 1.7490086693064326, "learning_rate": 5.000000000000021e-08, "loss": 2.2078, "step": 3995 }, { "epoch": 3.8948393378773125, "grad_norm": 1.8272955450221657, "learning_rate": 5.000000000000013e-08, "loss": 2.2437, "step": 4000 }, { "epoch": 3.8948393378773125, "eval_loss": 2.4291422367095947, "eval_runtime": 85.7591, "eval_samples_per_second": 85.111, "eval_steps_per_second": 0.676, "step": 4000 }, { "epoch": 3.8997078870496593, "grad_norm": 1.7590084186341215, "learning_rate": 5.000000000000008e-08, "loss": 2.2069, "step": 4005 }, { "epoch": 3.9045764362220057, "grad_norm": 1.7702717336824394, "learning_rate": 5.0000000000000044e-08, "loss": 2.1958, "step": 4010 }, { "epoch": 3.9094449853943525, "grad_norm": 1.7719070457541741, "learning_rate": 5.0000000000000024e-08, "loss": 2.1862, "step": 4015 }, { "epoch": 3.9143135345666993, "grad_norm": 1.729790978044211, "learning_rate": 5.000000000000001e-08, "loss": 2.201, "step": 4020 }, { "epoch": 3.9191820837390456, "grad_norm": 1.751806658583263, "learning_rate": 5.0000000000000004e-08, "loss": 2.2071, "step": 4025 }, { "epoch": 3.9240506329113924, "grad_norm": 1.7425275946141223, "learning_rate": 5.0000000000000004e-08, "loss": 2.2775, "step": 4030 }, { "epoch": 3.928919182083739, "grad_norm": 1.792394595138285, "learning_rate": 5e-08, "loss": 2.1683, "step": 4035 }, { "epoch": 3.9337877312560856, "grad_norm": 1.7949487182841233, "learning_rate": 5e-08, "loss": 2.1962, "step": 4040 }, { "epoch": 3.9386562804284324, "grad_norm": 1.8476999521081328, "learning_rate": 5e-08, "loss": 2.2004, "step": 4045 }, { "epoch": 3.943524829600779, "grad_norm": 1.8225182124751913, "learning_rate": 5e-08, "loss": 2.2092, "step": 4050 }, { "epoch": 3.9483933787731256, "grad_norm": 1.819449204564278, "learning_rate": 5e-08, "loss": 2.2543, "step": 4055 }, { "epoch": 3.9532619279454724, "grad_norm": 1.8290111060640115, "learning_rate": 5e-08, "loss": 2.1963, "step": 4060 }, { "epoch": 3.9581304771178187, "grad_norm": 1.7296724483111061, "learning_rate": 5e-08, "loss": 2.2167, "step": 4065 }, { "epoch": 3.9629990262901655, "grad_norm": 1.768342281901305, "learning_rate": 5e-08, "loss": 2.2145, "step": 4070 }, { "epoch": 3.9678675754625123, "grad_norm": 1.7810841025261306, "learning_rate": 5e-08, "loss": 2.2205, "step": 4075 }, { "epoch": 3.9727361246348587, "grad_norm": 1.7724437878330235, "learning_rate": 5e-08, "loss": 2.2227, "step": 4080 }, { "epoch": 3.9776046738072055, "grad_norm": 1.7693706025925438, "learning_rate": 5e-08, "loss": 2.26, "step": 4085 }, { "epoch": 3.982473222979552, "grad_norm": 1.8547893721966033, "learning_rate": 5e-08, "loss": 2.1705, "step": 4090 }, { "epoch": 3.9873417721518987, "grad_norm": 1.7856750671515411, "learning_rate": 5e-08, "loss": 2.2304, "step": 4095 }, { "epoch": 3.9922103213242455, "grad_norm": 1.789882341208583, "learning_rate": 5e-08, "loss": 2.2391, "step": 4100 }, { "epoch": 3.9970788704965923, "grad_norm": 1.8424154832009545, "learning_rate": 5e-08, "loss": 2.2157, "step": 4105 }, { "epoch": 4.0, "step": 4108, "total_flos": 429332983971840.0, "train_loss": 2.284057284167586, "train_runtime": 15107.0718, "train_samples_per_second": 17.393, "train_steps_per_second": 0.272 } ], "logging_steps": 5, "max_steps": 4108, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 200, "total_flos": 429332983971840.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }