{ "best_metric": 0.6744288802146912, "best_model_checkpoint": "models/llama-3.2-3b-sft-2/checkpoint-10000", "epoch": 1.3452613170108294, "eval_steps": 500, "global_step": 10000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00013452613170108295, "grad_norm": 5.089283466339111, "learning_rate": 5e-09, "loss": 1.0115, "step": 1 }, { "epoch": 0.006726306585054147, "grad_norm": 3.6531453132629395, "learning_rate": 2.5000000000000004e-07, "loss": 1.0329, "step": 50 }, { "epoch": 0.013452613170108294, "grad_norm": 1.954504132270813, "learning_rate": 5.000000000000001e-07, "loss": 0.9628, "step": 100 }, { "epoch": 0.02017891975516244, "grad_norm": 1.6202716827392578, "learning_rate": 7.5e-07, "loss": 0.882, "step": 150 }, { "epoch": 0.02690522634021659, "grad_norm": 1.5833114385604858, "learning_rate": 1.0000000000000002e-06, "loss": 0.8489, "step": 200 }, { "epoch": 0.033631532925270734, "grad_norm": 1.6729789972305298, "learning_rate": 1.25e-06, "loss": 0.8464, "step": 250 }, { "epoch": 0.04035783951032488, "grad_norm": 1.551665186882019, "learning_rate": 1.5e-06, "loss": 0.8205, "step": 300 }, { "epoch": 0.047084146095379024, "grad_norm": 1.6382861137390137, "learning_rate": 1.75e-06, "loss": 0.7994, "step": 350 }, { "epoch": 0.05381045268043318, "grad_norm": 1.8948196172714233, "learning_rate": 2.0000000000000003e-06, "loss": 0.7942, "step": 400 }, { "epoch": 0.06053675926548732, "grad_norm": 1.5274723768234253, "learning_rate": 2.25e-06, "loss": 0.7843, "step": 450 }, { "epoch": 0.06726306585054147, "grad_norm": 2.083868980407715, "learning_rate": 2.5e-06, "loss": 0.7792, "step": 500 }, { "epoch": 0.06726306585054147, "eval_loss": 0.772584855556488, "eval_runtime": 412.1529, "eval_samples_per_second": 121.501, "eval_steps_per_second": 1.9, "step": 500 }, { "epoch": 0.07398937243559561, "grad_norm": 1.4476754665374756, "learning_rate": 2.7500000000000004e-06, "loss": 0.7598, "step": 550 }, { "epoch": 0.08071567902064976, "grad_norm": 1.5177018642425537, "learning_rate": 3e-06, "loss": 0.7672, "step": 600 }, { "epoch": 0.0874419856057039, "grad_norm": 1.5025829076766968, "learning_rate": 3.2500000000000002e-06, "loss": 0.7442, "step": 650 }, { "epoch": 0.09416829219075805, "grad_norm": 1.6595162153244019, "learning_rate": 3.5e-06, "loss": 0.7562, "step": 700 }, { "epoch": 0.10089459877581221, "grad_norm": 1.4664040803909302, "learning_rate": 3.7500000000000005e-06, "loss": 0.7485, "step": 750 }, { "epoch": 0.10762090536086635, "grad_norm": 1.5404261350631714, "learning_rate": 4.000000000000001e-06, "loss": 0.7482, "step": 800 }, { "epoch": 0.1143472119459205, "grad_norm": 1.553759217262268, "learning_rate": 4.25e-06, "loss": 0.7475, "step": 850 }, { "epoch": 0.12107351853097464, "grad_norm": 1.5157835483551025, "learning_rate": 4.5e-06, "loss": 0.7403, "step": 900 }, { "epoch": 0.12779982511602878, "grad_norm": 1.4276628494262695, "learning_rate": 4.75e-06, "loss": 0.7392, "step": 950 }, { "epoch": 0.13452613170108293, "grad_norm": 1.5712127685546875, "learning_rate": 5e-06, "loss": 0.7496, "step": 1000 }, { "epoch": 0.13452613170108293, "eval_loss": 0.7444220185279846, "eval_runtime": 412.2013, "eval_samples_per_second": 121.487, "eval_steps_per_second": 1.9, "step": 1000 }, { "epoch": 0.1412524382861371, "grad_norm": 1.3969111442565918, "learning_rate": 4.9996192378909785e-06, "loss": 0.7548, "step": 1050 }, { "epoch": 0.14797874487119123, "grad_norm": 1.4484981298446655, "learning_rate": 4.99847706754774e-06, "loss": 0.73, "step": 1100 }, { "epoch": 0.15470505145624538, "grad_norm": 1.4184221029281616, "learning_rate": 4.9965738368864345e-06, "loss": 0.734, "step": 1150 }, { "epoch": 0.16143135804129952, "grad_norm": 1.4049128293991089, "learning_rate": 4.993910125649561e-06, "loss": 0.7295, "step": 1200 }, { "epoch": 0.16815766462635368, "grad_norm": 1.3905514478683472, "learning_rate": 4.990486745229364e-06, "loss": 0.7405, "step": 1250 }, { "epoch": 0.1748839712114078, "grad_norm": 1.5445802211761475, "learning_rate": 4.986304738420684e-06, "loss": 0.7198, "step": 1300 }, { "epoch": 0.18161027779646197, "grad_norm": 1.4158939123153687, "learning_rate": 4.981365379103306e-06, "loss": 0.7355, "step": 1350 }, { "epoch": 0.1883365843815161, "grad_norm": 1.3896397352218628, "learning_rate": 4.975670171853926e-06, "loss": 0.7168, "step": 1400 }, { "epoch": 0.19506289096657026, "grad_norm": 1.5642178058624268, "learning_rate": 4.9692208514878445e-06, "loss": 0.7369, "step": 1450 }, { "epoch": 0.20178919755162442, "grad_norm": 1.287514328956604, "learning_rate": 4.962019382530521e-06, "loss": 0.7243, "step": 1500 }, { "epoch": 0.20178919755162442, "eval_loss": 0.7295696139335632, "eval_runtime": 408.0925, "eval_samples_per_second": 122.71, "eval_steps_per_second": 1.919, "step": 1500 }, { "epoch": 0.20851550413667855, "grad_norm": 1.380998969078064, "learning_rate": 4.9540679586191605e-06, "loss": 0.7309, "step": 1550 }, { "epoch": 0.2152418107217327, "grad_norm": 1.4188984632492065, "learning_rate": 4.9453690018345144e-06, "loss": 0.7217, "step": 1600 }, { "epoch": 0.22196811730678684, "grad_norm": 1.306766390800476, "learning_rate": 4.935925161963089e-06, "loss": 0.7336, "step": 1650 }, { "epoch": 0.228694423891841, "grad_norm": 1.460458517074585, "learning_rate": 4.925739315689991e-06, "loss": 0.7258, "step": 1700 }, { "epoch": 0.23542073047689513, "grad_norm": 1.4665378332138062, "learning_rate": 4.914814565722671e-06, "loss": 0.7249, "step": 1750 }, { "epoch": 0.2421470370619493, "grad_norm": 1.5657004117965698, "learning_rate": 4.903154239845798e-06, "loss": 0.7253, "step": 1800 }, { "epoch": 0.24887334364700342, "grad_norm": 1.3820271492004395, "learning_rate": 4.890761889907589e-06, "loss": 0.7151, "step": 1850 }, { "epoch": 0.25559965023205755, "grad_norm": 1.4721177816390991, "learning_rate": 4.8776412907378845e-06, "loss": 0.7251, "step": 1900 }, { "epoch": 0.26232595681711174, "grad_norm": 1.4119197130203247, "learning_rate": 4.863796438998293e-06, "loss": 0.7188, "step": 1950 }, { "epoch": 0.26905226340216587, "grad_norm": 1.3871151208877563, "learning_rate": 4.849231551964771e-06, "loss": 0.7178, "step": 2000 }, { "epoch": 0.26905226340216587, "eval_loss": 0.7197288870811462, "eval_runtime": 407.583, "eval_samples_per_second": 122.863, "eval_steps_per_second": 1.921, "step": 2000 }, { "epoch": 0.27577856998722, "grad_norm": 1.4082683324813843, "learning_rate": 4.833951066243004e-06, "loss": 0.7255, "step": 2050 }, { "epoch": 0.2825048765722742, "grad_norm": 1.3477662801742554, "learning_rate": 4.817959636416969e-06, "loss": 0.7141, "step": 2100 }, { "epoch": 0.2892311831573283, "grad_norm": 1.316319227218628, "learning_rate": 4.801262133631101e-06, "loss": 0.7146, "step": 2150 }, { "epoch": 0.29595748974238245, "grad_norm": 1.3264553546905518, "learning_rate": 4.783863644106502e-06, "loss": 0.7176, "step": 2200 }, { "epoch": 0.3026837963274366, "grad_norm": 1.4020483493804932, "learning_rate": 4.765769467591626e-06, "loss": 0.7082, "step": 2250 }, { "epoch": 0.30941010291249077, "grad_norm": 1.4155000448226929, "learning_rate": 4.746985115747918e-06, "loss": 0.7133, "step": 2300 }, { "epoch": 0.3161364094975449, "grad_norm": 1.3636876344680786, "learning_rate": 4.72751631047092e-06, "loss": 0.721, "step": 2350 }, { "epoch": 0.32286271608259903, "grad_norm": 1.3558250665664673, "learning_rate": 4.707368982147318e-06, "loss": 0.7056, "step": 2400 }, { "epoch": 0.32958902266765316, "grad_norm": 1.3763400316238403, "learning_rate": 4.68654926784849e-06, "loss": 0.7157, "step": 2450 }, { "epoch": 0.33631532925270735, "grad_norm": 1.4673148393630981, "learning_rate": 4.665063509461098e-06, "loss": 0.7077, "step": 2500 }, { "epoch": 0.33631532925270735, "eval_loss": 0.7126947045326233, "eval_runtime": 407.7534, "eval_samples_per_second": 122.812, "eval_steps_per_second": 1.92, "step": 2500 }, { "epoch": 0.3430416358377615, "grad_norm": 1.5427167415618896, "learning_rate": 4.642918251755281e-06, "loss": 0.7216, "step": 2550 }, { "epoch": 0.3497679424228156, "grad_norm": 1.4927399158477783, "learning_rate": 4.620120240391065e-06, "loss": 0.7096, "step": 2600 }, { "epoch": 0.3564942490078698, "grad_norm": 1.3618338108062744, "learning_rate": 4.596676419863561e-06, "loss": 0.7106, "step": 2650 }, { "epoch": 0.36322055559292393, "grad_norm": 1.3076339960098267, "learning_rate": 4.572593931387604e-06, "loss": 0.7213, "step": 2700 }, { "epoch": 0.36994686217797806, "grad_norm": 1.2195158004760742, "learning_rate": 4.54788011072248e-06, "loss": 0.7102, "step": 2750 }, { "epoch": 0.3766731687630322, "grad_norm": 1.341899037361145, "learning_rate": 4.522542485937369e-06, "loss": 0.7021, "step": 2800 }, { "epoch": 0.3833994753480864, "grad_norm": 1.4170920848846436, "learning_rate": 4.496588775118232e-06, "loss": 0.6942, "step": 2850 }, { "epoch": 0.3901257819331405, "grad_norm": 1.502806305885315, "learning_rate": 4.470026884016805e-06, "loss": 0.6947, "step": 2900 }, { "epoch": 0.39685208851819465, "grad_norm": 1.3311142921447754, "learning_rate": 4.442864903642428e-06, "loss": 0.6926, "step": 2950 }, { "epoch": 0.40357839510324883, "grad_norm": 1.3881827592849731, "learning_rate": 4.415111107797445e-06, "loss": 0.6992, "step": 3000 }, { "epoch": 0.40357839510324883, "eval_loss": 0.7066433429718018, "eval_runtime": 407.4989, "eval_samples_per_second": 122.889, "eval_steps_per_second": 1.921, "step": 3000 }, { "epoch": 0.41030470168830296, "grad_norm": 1.4951732158660889, "learning_rate": 4.386773950556931e-06, "loss": 0.6967, "step": 3050 }, { "epoch": 0.4170310082733571, "grad_norm": 1.3532462120056152, "learning_rate": 4.357862063693486e-06, "loss": 0.7056, "step": 3100 }, { "epoch": 0.4237573148584112, "grad_norm": 1.3927810192108154, "learning_rate": 4.328384254047927e-06, "loss": 0.6962, "step": 3150 }, { "epoch": 0.4304836214434654, "grad_norm": 1.2891749143600464, "learning_rate": 4.2983495008466285e-06, "loss": 0.7193, "step": 3200 }, { "epoch": 0.43720992802851955, "grad_norm": 1.3432807922363281, "learning_rate": 4.267766952966369e-06, "loss": 0.7008, "step": 3250 }, { "epoch": 0.4439362346135737, "grad_norm": 1.5565185546875, "learning_rate": 4.236645926147493e-06, "loss": 0.6939, "step": 3300 }, { "epoch": 0.4506625411986278, "grad_norm": 1.3859912157058716, "learning_rate": 4.204995900156247e-06, "loss": 0.6953, "step": 3350 }, { "epoch": 0.457388847783682, "grad_norm": 1.358071208000183, "learning_rate": 4.172826515897146e-06, "loss": 0.6847, "step": 3400 }, { "epoch": 0.4641151543687361, "grad_norm": 1.3290272951126099, "learning_rate": 4.140147572476269e-06, "loss": 0.7069, "step": 3450 }, { "epoch": 0.47084146095379026, "grad_norm": 1.387556552886963, "learning_rate": 4.106969024216348e-06, "loss": 0.6992, "step": 3500 }, { "epoch": 0.47084146095379026, "eval_loss": 0.7012136578559875, "eval_runtime": 407.5913, "eval_samples_per_second": 122.861, "eval_steps_per_second": 1.921, "step": 3500 }, { "epoch": 0.47756776753884445, "grad_norm": 1.4065874814987183, "learning_rate": 4.073300977624594e-06, "loss": 0.7044, "step": 3550 }, { "epoch": 0.4842940741238986, "grad_norm": 1.3837950229644775, "learning_rate": 4.039153688314146e-06, "loss": 0.7002, "step": 3600 }, { "epoch": 0.4910203807089527, "grad_norm": 1.345460057258606, "learning_rate": 4.0045375578801216e-06, "loss": 0.692, "step": 3650 }, { "epoch": 0.49774668729400684, "grad_norm": 1.3991773128509521, "learning_rate": 3.969463130731183e-06, "loss": 0.691, "step": 3700 }, { "epoch": 0.504472993879061, "grad_norm": 1.4494290351867676, "learning_rate": 3.933941090877615e-06, "loss": 0.6975, "step": 3750 }, { "epoch": 0.5111993004641151, "grad_norm": 1.4690207242965698, "learning_rate": 3.897982258676867e-06, "loss": 0.6951, "step": 3800 }, { "epoch": 0.5179256070491693, "grad_norm": 1.2572777271270752, "learning_rate": 3.861597587537568e-06, "loss": 0.6997, "step": 3850 }, { "epoch": 0.5246519136342235, "grad_norm": 1.416960597038269, "learning_rate": 3.824798160583012e-06, "loss": 0.6883, "step": 3900 }, { "epoch": 0.5313782202192776, "grad_norm": 1.3925755023956299, "learning_rate": 3.787595187275136e-06, "loss": 0.6894, "step": 3950 }, { "epoch": 0.5381045268043317, "grad_norm": 1.2715755701065063, "learning_rate": 3.7500000000000005e-06, "loss": 0.6945, "step": 4000 }, { "epoch": 0.5381045268043317, "eval_loss": 0.6965380907058716, "eval_runtime": 407.4593, "eval_samples_per_second": 122.901, "eval_steps_per_second": 1.922, "step": 4000 }, { "epoch": 0.5448308333893859, "grad_norm": 1.1935741901397705, "learning_rate": 3.7120240506158433e-06, "loss": 0.6899, "step": 4050 }, { "epoch": 0.55155713997444, "grad_norm": 1.3253408670425415, "learning_rate": 3.6736789069647273e-06, "loss": 0.6964, "step": 4100 }, { "epoch": 0.5582834465594941, "grad_norm": 1.4265397787094116, "learning_rate": 3.634976249348867e-06, "loss": 0.6867, "step": 4150 }, { "epoch": 0.5650097531445484, "grad_norm": 1.3048572540283203, "learning_rate": 3.595927866972694e-06, "loss": 0.6959, "step": 4200 }, { "epoch": 0.5717360597296025, "grad_norm": 1.495254397392273, "learning_rate": 3.556545654351749e-06, "loss": 0.6841, "step": 4250 }, { "epoch": 0.5784623663146566, "grad_norm": 1.3568543195724487, "learning_rate": 3.516841607689501e-06, "loss": 0.6911, "step": 4300 }, { "epoch": 0.5851886728997108, "grad_norm": 1.274553656578064, "learning_rate": 3.476827821223184e-06, "loss": 0.6819, "step": 4350 }, { "epoch": 0.5919149794847649, "grad_norm": 1.3335059881210327, "learning_rate": 3.436516483539781e-06, "loss": 0.7018, "step": 4400 }, { "epoch": 0.598641286069819, "grad_norm": 1.265262246131897, "learning_rate": 3.39591987386325e-06, "loss": 0.6821, "step": 4450 }, { "epoch": 0.6053675926548732, "grad_norm": 1.3332948684692383, "learning_rate": 3.3550503583141726e-06, "loss": 0.6879, "step": 4500 }, { "epoch": 0.6053675926548732, "eval_loss": 0.6919859647750854, "eval_runtime": 407.9173, "eval_samples_per_second": 122.763, "eval_steps_per_second": 1.92, "step": 4500 }, { "epoch": 0.6120938992399274, "grad_norm": 1.310651421546936, "learning_rate": 3.313920386142892e-06, "loss": 0.6857, "step": 4550 }, { "epoch": 0.6188202058249815, "grad_norm": 1.2892873287200928, "learning_rate": 3.272542485937369e-06, "loss": 0.6812, "step": 4600 }, { "epoch": 0.6255465124100357, "grad_norm": 1.2226920127868652, "learning_rate": 3.230929261806842e-06, "loss": 0.6774, "step": 4650 }, { "epoch": 0.6322728189950898, "grad_norm": 1.2619507312774658, "learning_rate": 3.189093389542498e-06, "loss": 0.6832, "step": 4700 }, { "epoch": 0.6389991255801439, "grad_norm": 1.4155713319778442, "learning_rate": 3.147047612756302e-06, "loss": 0.6897, "step": 4750 }, { "epoch": 0.6457254321651981, "grad_norm": 1.3170417547225952, "learning_rate": 3.1048047389991693e-06, "loss": 0.6962, "step": 4800 }, { "epoch": 0.6524517387502522, "grad_norm": 1.3795112371444702, "learning_rate": 3.062377635859663e-06, "loss": 0.6838, "step": 4850 }, { "epoch": 0.6591780453353063, "grad_norm": 1.3435876369476318, "learning_rate": 3.019779227044398e-06, "loss": 0.6813, "step": 4900 }, { "epoch": 0.6659043519203606, "grad_norm": 1.4538263082504272, "learning_rate": 2.9770224884413625e-06, "loss": 0.6941, "step": 4950 }, { "epoch": 0.6726306585054147, "grad_norm": 1.4418572187423706, "learning_rate": 2.9341204441673267e-06, "loss": 0.6901, "step": 5000 }, { "epoch": 0.6726306585054147, "eval_loss": 0.6878845691680908, "eval_runtime": 407.8759, "eval_samples_per_second": 122.775, "eval_steps_per_second": 1.92, "step": 5000 }, { "epoch": 0.6793569650904688, "grad_norm": 1.3608489036560059, "learning_rate": 2.8910861626005774e-06, "loss": 0.6822, "step": 5050 }, { "epoch": 0.686083271675523, "grad_norm": 1.2994650602340698, "learning_rate": 2.847932752400164e-06, "loss": 0.6762, "step": 5100 }, { "epoch": 0.6928095782605771, "grad_norm": 1.565314531326294, "learning_rate": 2.804673358512869e-06, "loss": 0.686, "step": 5150 }, { "epoch": 0.6995358848456312, "grad_norm": 1.3825584650039673, "learning_rate": 2.761321158169134e-06, "loss": 0.6872, "step": 5200 }, { "epoch": 0.7062621914306854, "grad_norm": 1.3237903118133545, "learning_rate": 2.717889356869146e-06, "loss": 0.6873, "step": 5250 }, { "epoch": 0.7129884980157396, "grad_norm": 1.4130380153656006, "learning_rate": 2.6743911843603134e-06, "loss": 0.6849, "step": 5300 }, { "epoch": 0.7197148046007937, "grad_norm": 1.3584030866622925, "learning_rate": 2.6308398906073603e-06, "loss": 0.6914, "step": 5350 }, { "epoch": 0.7264411111858479, "grad_norm": 1.4434551000595093, "learning_rate": 2.587248741756253e-06, "loss": 0.6815, "step": 5400 }, { "epoch": 0.733167417770902, "grad_norm": 1.343894362449646, "learning_rate": 2.543631016093209e-06, "loss": 0.6739, "step": 5450 }, { "epoch": 0.7398937243559561, "grad_norm": 1.5112074613571167, "learning_rate": 2.5e-06, "loss": 0.6759, "step": 5500 }, { "epoch": 0.7398937243559561, "eval_loss": 0.6843944787979126, "eval_runtime": 407.3845, "eval_samples_per_second": 122.923, "eval_steps_per_second": 1.922, "step": 5500 }, { "epoch": 0.7466200309410103, "grad_norm": 1.471646785736084, "learning_rate": 2.4563689839067913e-06, "loss": 0.6914, "step": 5550 }, { "epoch": 0.7533463375260644, "grad_norm": 1.3042908906936646, "learning_rate": 2.4127512582437486e-06, "loss": 0.6813, "step": 5600 }, { "epoch": 0.7600726441111186, "grad_norm": 1.2429543733596802, "learning_rate": 2.3691601093926406e-06, "loss": 0.6757, "step": 5650 }, { "epoch": 0.7667989506961728, "grad_norm": 1.300581932067871, "learning_rate": 2.325608815639687e-06, "loss": 0.6861, "step": 5700 }, { "epoch": 0.7735252572812269, "grad_norm": 1.3551619052886963, "learning_rate": 2.2821106431308546e-06, "loss": 0.677, "step": 5750 }, { "epoch": 0.780251563866281, "grad_norm": 1.2355047464370728, "learning_rate": 2.238678841830867e-06, "loss": 0.6728, "step": 5800 }, { "epoch": 0.7869778704513352, "grad_norm": 1.3812388181686401, "learning_rate": 2.195326641487132e-06, "loss": 0.6807, "step": 5850 }, { "epoch": 0.7937041770363893, "grad_norm": 1.3730298280715942, "learning_rate": 2.1520672475998374e-06, "loss": 0.6835, "step": 5900 }, { "epoch": 0.8004304836214434, "grad_norm": 1.3047642707824707, "learning_rate": 2.1089138373994226e-06, "loss": 0.695, "step": 5950 }, { "epoch": 0.8071567902064977, "grad_norm": 1.2433747053146362, "learning_rate": 2.0658795558326745e-06, "loss": 0.6752, "step": 6000 }, { "epoch": 0.8071567902064977, "eval_loss": 0.681221604347229, "eval_runtime": 407.827, "eval_samples_per_second": 122.79, "eval_steps_per_second": 1.92, "step": 6000 }, { "epoch": 0.8138830967915518, "grad_norm": 1.245276927947998, "learning_rate": 2.022977511558638e-06, "loss": 0.6822, "step": 6050 }, { "epoch": 0.8206094033766059, "grad_norm": 1.306039810180664, "learning_rate": 1.9802207729556023e-06, "loss": 0.6902, "step": 6100 }, { "epoch": 0.8273357099616601, "grad_norm": 1.3029533624649048, "learning_rate": 1.937622364140338e-06, "loss": 0.69, "step": 6150 }, { "epoch": 0.8340620165467142, "grad_norm": 1.2798274755477905, "learning_rate": 1.895195261000831e-06, "loss": 0.6791, "step": 6200 }, { "epoch": 0.8407883231317683, "grad_norm": 1.3335996866226196, "learning_rate": 1.852952387243698e-06, "loss": 0.6807, "step": 6250 }, { "epoch": 0.8475146297168225, "grad_norm": 1.3937443494796753, "learning_rate": 1.8109066104575023e-06, "loss": 0.6817, "step": 6300 }, { "epoch": 0.8542409363018766, "grad_norm": 1.3417977094650269, "learning_rate": 1.7690707381931585e-06, "loss": 0.6824, "step": 6350 }, { "epoch": 0.8609672428869308, "grad_norm": 1.2701561450958252, "learning_rate": 1.7274575140626318e-06, "loss": 0.674, "step": 6400 }, { "epoch": 0.867693549471985, "grad_norm": 1.2082581520080566, "learning_rate": 1.686079613857109e-06, "loss": 0.6725, "step": 6450 }, { "epoch": 0.8744198560570391, "grad_norm": 1.4414860010147095, "learning_rate": 1.6449496416858285e-06, "loss": 0.6826, "step": 6500 }, { "epoch": 0.8744198560570391, "eval_loss": 0.6783373951911926, "eval_runtime": 407.3438, "eval_samples_per_second": 122.935, "eval_steps_per_second": 1.922, "step": 6500 }, { "epoch": 0.8811461626420932, "grad_norm": 1.2324413061141968, "learning_rate": 1.6040801261367494e-06, "loss": 0.6682, "step": 6550 }, { "epoch": 0.8878724692271474, "grad_norm": 1.2953166961669922, "learning_rate": 1.56348351646022e-06, "loss": 0.6811, "step": 6600 }, { "epoch": 0.8945987758122015, "grad_norm": 1.2646501064300537, "learning_rate": 1.5231721787768162e-06, "loss": 0.6813, "step": 6650 }, { "epoch": 0.9013250823972556, "grad_norm": 1.4454089403152466, "learning_rate": 1.4831583923105e-06, "loss": 0.6744, "step": 6700 }, { "epoch": 0.9080513889823099, "grad_norm": 1.3868876695632935, "learning_rate": 1.443454345648252e-06, "loss": 0.6764, "step": 6750 }, { "epoch": 0.914777695567364, "grad_norm": 1.2375882863998413, "learning_rate": 1.4040721330273063e-06, "loss": 0.678, "step": 6800 }, { "epoch": 0.9215040021524181, "grad_norm": 1.3029659986495972, "learning_rate": 1.3650237506511333e-06, "loss": 0.6736, "step": 6850 }, { "epoch": 0.9282303087374723, "grad_norm": 1.3237438201904297, "learning_rate": 1.3263210930352737e-06, "loss": 0.6769, "step": 6900 }, { "epoch": 0.9349566153225264, "grad_norm": 1.3647958040237427, "learning_rate": 1.2879759493841577e-06, "loss": 0.6663, "step": 6950 }, { "epoch": 0.9416829219075805, "grad_norm": 1.2964695692062378, "learning_rate": 1.2500000000000007e-06, "loss": 0.6804, "step": 7000 }, { "epoch": 0.9416829219075805, "eval_loss": 0.6757956147193909, "eval_runtime": 407.4829, "eval_samples_per_second": 122.894, "eval_steps_per_second": 1.922, "step": 7000 }, { "epoch": 0.9484092284926346, "grad_norm": 1.444977045059204, "learning_rate": 1.2124048127248644e-06, "loss": 0.6816, "step": 7050 }, { "epoch": 0.9551355350776889, "grad_norm": 1.262220025062561, "learning_rate": 1.1752018394169882e-06, "loss": 0.6687, "step": 7100 }, { "epoch": 0.961861841662743, "grad_norm": 1.2450040578842163, "learning_rate": 1.1384024124624324e-06, "loss": 0.6738, "step": 7150 }, { "epoch": 0.9685881482477972, "grad_norm": 1.31246817111969, "learning_rate": 1.1020177413231334e-06, "loss": 0.6664, "step": 7200 }, { "epoch": 0.9753144548328513, "grad_norm": 1.307934284210205, "learning_rate": 1.0660589091223854e-06, "loss": 0.6845, "step": 7250 }, { "epoch": 0.9820407614179054, "grad_norm": 1.2466742992401123, "learning_rate": 1.0305368692688175e-06, "loss": 0.6732, "step": 7300 }, { "epoch": 0.9887670680029595, "grad_norm": 1.3563283681869507, "learning_rate": 9.95462442119879e-07, "loss": 0.6683, "step": 7350 }, { "epoch": 0.9954933745880137, "grad_norm": 1.344889760017395, "learning_rate": 9.608463116858544e-07, "loss": 0.6794, "step": 7400 }, { "epoch": 1.002219681173068, "grad_norm": 1.2191548347473145, "learning_rate": 9.266990223754069e-07, "loss": 0.6551, "step": 7450 }, { "epoch": 1.008945987758122, "grad_norm": 1.4573768377304077, "learning_rate": 8.930309757836517e-07, "loss": 0.6131, "step": 7500 }, { "epoch": 1.008945987758122, "eval_loss": 0.6763768196105957, "eval_runtime": 407.3062, "eval_samples_per_second": 122.947, "eval_steps_per_second": 1.922, "step": 7500 }, { "epoch": 1.0156722943431762, "grad_norm": 1.4787367582321167, "learning_rate": 8.598524275237321e-07, "loss": 0.5996, "step": 7550 }, { "epoch": 1.0223986009282302, "grad_norm": 1.2250146865844727, "learning_rate": 8.271734841028553e-07, "loss": 0.6111, "step": 7600 }, { "epoch": 1.0291249075132844, "grad_norm": 1.2959485054016113, "learning_rate": 7.950040998437541e-07, "loss": 0.5994, "step": 7650 }, { "epoch": 1.0358512140983387, "grad_norm": 1.3388257026672363, "learning_rate": 7.633540738525066e-07, "loss": 0.6092, "step": 7700 }, { "epoch": 1.0425775206833927, "grad_norm": 1.3289120197296143, "learning_rate": 7.322330470336314e-07, "loss": 0.6154, "step": 7750 }, { "epoch": 1.049303827268447, "grad_norm": 1.4141819477081299, "learning_rate": 7.016504991533727e-07, "loss": 0.6132, "step": 7800 }, { "epoch": 1.056030133853501, "grad_norm": 1.368107557296753, "learning_rate": 6.716157459520739e-07, "loss": 0.6147, "step": 7850 }, { "epoch": 1.0627564404385552, "grad_norm": 1.3877215385437012, "learning_rate": 6.421379363065142e-07, "loss": 0.6234, "step": 7900 }, { "epoch": 1.0694827470236092, "grad_norm": 1.3689961433410645, "learning_rate": 6.1322604944307e-07, "loss": 0.6041, "step": 7950 }, { "epoch": 1.0762090536086635, "grad_norm": 1.4440808296203613, "learning_rate": 5.848888922025553e-07, "loss": 0.6012, "step": 8000 }, { "epoch": 1.0762090536086635, "eval_loss": 0.6758261322975159, "eval_runtime": 407.4199, "eval_samples_per_second": 122.912, "eval_steps_per_second": 1.922, "step": 8000 }, { "epoch": 1.0829353601937177, "grad_norm": 1.3043277263641357, "learning_rate": 5.571350963575728e-07, "loss": 0.6173, "step": 8050 }, { "epoch": 1.0896616667787717, "grad_norm": 1.3491568565368652, "learning_rate": 5.299731159831953e-07, "loss": 0.6101, "step": 8100 }, { "epoch": 1.096387973363826, "grad_norm": 1.3337068557739258, "learning_rate": 5.034112248817685e-07, "loss": 0.6157, "step": 8150 }, { "epoch": 1.10311427994888, "grad_norm": 1.4220951795578003, "learning_rate": 4.774575140626317e-07, "loss": 0.613, "step": 8200 }, { "epoch": 1.1098405865339342, "grad_norm": 1.3012259006500244, "learning_rate": 4.5211988927752026e-07, "loss": 0.6025, "step": 8250 }, { "epoch": 1.1165668931189883, "grad_norm": 1.3978270292282104, "learning_rate": 4.27406068612396e-07, "loss": 0.6086, "step": 8300 }, { "epoch": 1.1232931997040425, "grad_norm": 1.381862998008728, "learning_rate": 4.033235801364402e-07, "loss": 0.6044, "step": 8350 }, { "epoch": 1.1300195062890968, "grad_norm": 1.3832786083221436, "learning_rate": 3.798797596089351e-07, "loss": 0.6117, "step": 8400 }, { "epoch": 1.1367458128741508, "grad_norm": 1.3328065872192383, "learning_rate": 3.5708174824471947e-07, "loss": 0.6021, "step": 8450 }, { "epoch": 1.143472119459205, "grad_norm": 1.5573444366455078, "learning_rate": 3.3493649053890325e-07, "loss": 0.6136, "step": 8500 }, { "epoch": 1.143472119459205, "eval_loss": 0.675076425075531, "eval_runtime": 407.3808, "eval_samples_per_second": 122.924, "eval_steps_per_second": 1.922, "step": 8500 }, { "epoch": 1.150198426044259, "grad_norm": 1.3546710014343262, "learning_rate": 3.134507321515107e-07, "loss": 0.6107, "step": 8550 }, { "epoch": 1.1569247326293133, "grad_norm": 1.2918901443481445, "learning_rate": 2.9263101785268253e-07, "loss": 0.6085, "step": 8600 }, { "epoch": 1.1636510392143673, "grad_norm": 1.3548731803894043, "learning_rate": 2.7248368952908055e-07, "loss": 0.6033, "step": 8650 }, { "epoch": 1.1703773457994215, "grad_norm": 1.4083459377288818, "learning_rate": 2.53014884252083e-07, "loss": 0.623, "step": 8700 }, { "epoch": 1.1771036523844756, "grad_norm": 1.3091658353805542, "learning_rate": 2.3423053240837518e-07, "loss": 0.6088, "step": 8750 }, { "epoch": 1.1838299589695298, "grad_norm": 1.394978404045105, "learning_rate": 2.1613635589349756e-07, "loss": 0.6069, "step": 8800 }, { "epoch": 1.190556265554584, "grad_norm": 1.3812838792800903, "learning_rate": 1.9873786636889908e-07, "loss": 0.6087, "step": 8850 }, { "epoch": 1.197282572139638, "grad_norm": 1.3317595720291138, "learning_rate": 1.8204036358303173e-07, "loss": 0.6059, "step": 8900 }, { "epoch": 1.2040088787246923, "grad_norm": 1.2979373931884766, "learning_rate": 1.6604893375699594e-07, "loss": 0.5978, "step": 8950 }, { "epoch": 1.2107351853097463, "grad_norm": 1.427206039428711, "learning_rate": 1.507684480352292e-07, "loss": 0.6127, "step": 9000 }, { "epoch": 1.2107351853097463, "eval_loss": 0.6747262477874756, "eval_runtime": 407.1101, "eval_samples_per_second": 123.006, "eval_steps_per_second": 1.923, "step": 9000 }, { "epoch": 1.2174614918948006, "grad_norm": 1.2737078666687012, "learning_rate": 1.362035610017079e-07, "loss": 0.6171, "step": 9050 }, { "epoch": 1.2241877984798548, "grad_norm": 1.4532371759414673, "learning_rate": 1.223587092621162e-07, "loss": 0.6093, "step": 9100 }, { "epoch": 1.2309141050649088, "grad_norm": 1.4686241149902344, "learning_rate": 1.0923811009241142e-07, "loss": 0.6073, "step": 9150 }, { "epoch": 1.237640411649963, "grad_norm": 1.3653959035873413, "learning_rate": 9.684576015420277e-08, "loss": 0.6131, "step": 9200 }, { "epoch": 1.244366718235017, "grad_norm": 1.3024178743362427, "learning_rate": 8.518543427732951e-08, "loss": 0.6162, "step": 9250 }, { "epoch": 1.2510930248200713, "grad_norm": 1.3427560329437256, "learning_rate": 7.426068431000883e-08, "loss": 0.6033, "step": 9300 }, { "epoch": 1.2578193314051254, "grad_norm": 1.4388611316680908, "learning_rate": 6.407483803691216e-08, "loss": 0.6104, "step": 9350 }, { "epoch": 1.2645456379901796, "grad_norm": 1.610689401626587, "learning_rate": 5.463099816548578e-08, "loss": 0.612, "step": 9400 }, { "epoch": 1.2712719445752336, "grad_norm": 1.4857087135314941, "learning_rate": 4.593204138084006e-08, "loss": 0.6069, "step": 9450 }, { "epoch": 1.2779982511602879, "grad_norm": 1.3649048805236816, "learning_rate": 3.798061746947995e-08, "loss": 0.6076, "step": 9500 }, { "epoch": 1.2779982511602879, "eval_loss": 0.6744668483734131, "eval_runtime": 407.221, "eval_samples_per_second": 122.973, "eval_steps_per_second": 1.923, "step": 9500 }, { "epoch": 1.284724557745342, "grad_norm": 1.2920268774032593, "learning_rate": 3.077914851215585e-08, "loss": 0.6195, "step": 9550 }, { "epoch": 1.2914508643303961, "grad_norm": 1.3456374406814575, "learning_rate": 2.4329828146074096e-08, "loss": 0.5986, "step": 9600 }, { "epoch": 1.2981771709154504, "grad_norm": 1.340410828590393, "learning_rate": 1.8634620896695044e-08, "loss": 0.6146, "step": 9650 }, { "epoch": 1.3049034775005044, "grad_norm": 1.4664082527160645, "learning_rate": 1.3695261579316776e-08, "loss": 0.6148, "step": 9700 }, { "epoch": 1.3116297840855586, "grad_norm": 1.3726171255111694, "learning_rate": 9.513254770636138e-09, "loss": 0.6112, "step": 9750 }, { "epoch": 1.3183560906706129, "grad_norm": 1.3762577772140503, "learning_rate": 6.089874350439507e-09, "loss": 0.6066, "step": 9800 }, { "epoch": 1.325082397255667, "grad_norm": 1.3914607763290405, "learning_rate": 3.4261631135654174e-09, "loss": 0.6114, "step": 9850 }, { "epoch": 1.3318087038407211, "grad_norm": 1.2674617767333984, "learning_rate": 1.5229324522605949e-09, "loss": 0.6165, "step": 9900 }, { "epoch": 1.3385350104257752, "grad_norm": 1.2815463542938232, "learning_rate": 3.8076210902182607e-10, "loss": 0.6101, "step": 9950 }, { "epoch": 1.3452613170108294, "grad_norm": 1.4005900621414185, "learning_rate": 0.0, "loss": 0.6033, "step": 10000 }, { "epoch": 1.3452613170108294, "eval_loss": 0.6744288802146912, "eval_runtime": 407.1493, "eval_samples_per_second": 122.994, "eval_steps_per_second": 1.923, "step": 10000 }, { "epoch": 1.3452613170108294, "step": 10000, "total_flos": 1119551134941184.0, "train_loss": 0.6844510423541069, "train_runtime": 79339.5992, "train_samples_per_second": 16.133, "train_steps_per_second": 0.126 } ], "logging_steps": 50, "max_steps": 10000, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1119551134941184.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }