{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 100, "global_step": 1231, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008123476848090982, "grad_norm": 3.203125, "learning_rate": 1.6129032258064516e-06, "loss": 1.5961, "step": 10 }, { "epoch": 0.016246953696181964, "grad_norm": 3.453125, "learning_rate": 3.225806451612903e-06, "loss": 1.5961, "step": 20 }, { "epoch": 0.024370430544272948, "grad_norm": 2.796875, "learning_rate": 4.838709677419355e-06, "loss": 1.5969, "step": 30 }, { "epoch": 0.03249390739236393, "grad_norm": 2.609375, "learning_rate": 6.451612903225806e-06, "loss": 1.5125, "step": 40 }, { "epoch": 0.04061738424045491, "grad_norm": 2.015625, "learning_rate": 8.064516129032258e-06, "loss": 1.4758, "step": 50 }, { "epoch": 0.048740861088545896, "grad_norm": 2.078125, "learning_rate": 9.67741935483871e-06, "loss": 1.3445, "step": 60 }, { "epoch": 0.05686433793663688, "grad_norm": 2.21875, "learning_rate": 1.1290322580645164e-05, "loss": 1.4297, "step": 70 }, { "epoch": 0.06498781478472786, "grad_norm": 1.84375, "learning_rate": 1.2903225806451613e-05, "loss": 1.3977, "step": 80 }, { "epoch": 0.07311129163281885, "grad_norm": 1.9765625, "learning_rate": 1.4516129032258066e-05, "loss": 1.3516, "step": 90 }, { "epoch": 0.08123476848090982, "grad_norm": 1.8828125, "learning_rate": 1.6129032258064517e-05, "loss": 1.3461, "step": 100 }, { "epoch": 0.08935824532900082, "grad_norm": 2.015625, "learning_rate": 1.774193548387097e-05, "loss": 1.3195, "step": 110 }, { "epoch": 0.09748172217709179, "grad_norm": 2.078125, "learning_rate": 1.935483870967742e-05, "loss": 1.3828, "step": 120 }, { "epoch": 0.10560519902518278, "grad_norm": 1.84375, "learning_rate": 1.989159891598916e-05, "loss": 1.3586, "step": 130 }, { "epoch": 0.11372867587327376, "grad_norm": 2.0625, "learning_rate": 1.971093044263776e-05, "loss": 1.3945, "step": 140 }, { "epoch": 0.12185215272136475, "grad_norm": 1.984375, "learning_rate": 1.9530261969286363e-05, "loss": 1.343, "step": 150 }, { "epoch": 0.12997562956945571, "grad_norm": 2.0625, "learning_rate": 1.934959349593496e-05, "loss": 1.3172, "step": 160 }, { "epoch": 0.13809910641754672, "grad_norm": 1.8671875, "learning_rate": 1.916892502258356e-05, "loss": 1.3141, "step": 170 }, { "epoch": 0.1462225832656377, "grad_norm": 2.171875, "learning_rate": 1.898825654923216e-05, "loss": 1.3242, "step": 180 }, { "epoch": 0.15434606011372867, "grad_norm": 1.921875, "learning_rate": 1.8807588075880762e-05, "loss": 1.3633, "step": 190 }, { "epoch": 0.16246953696181965, "grad_norm": 1.890625, "learning_rate": 1.862691960252936e-05, "loss": 1.3359, "step": 200 }, { "epoch": 0.17059301380991065, "grad_norm": 1.7421875, "learning_rate": 1.8446251129177958e-05, "loss": 1.3266, "step": 210 }, { "epoch": 0.17871649065800163, "grad_norm": 1.984375, "learning_rate": 1.826558265582656e-05, "loss": 1.3367, "step": 220 }, { "epoch": 0.1868399675060926, "grad_norm": 2.09375, "learning_rate": 1.808491418247516e-05, "loss": 1.3023, "step": 230 }, { "epoch": 0.19496344435418358, "grad_norm": 1.8515625, "learning_rate": 1.790424570912376e-05, "loss": 1.2961, "step": 240 }, { "epoch": 0.20308692120227456, "grad_norm": 1.859375, "learning_rate": 1.772357723577236e-05, "loss": 1.3711, "step": 250 }, { "epoch": 0.21121039805036557, "grad_norm": 1.9375, "learning_rate": 1.754290876242096e-05, "loss": 1.3062, "step": 260 }, { "epoch": 0.21933387489845654, "grad_norm": 2.140625, "learning_rate": 1.7362240289069557e-05, "loss": 1.3281, "step": 270 }, { "epoch": 0.22745735174654752, "grad_norm": 1.7734375, "learning_rate": 1.718157181571816e-05, "loss": 1.2883, "step": 280 }, { "epoch": 0.2355808285946385, "grad_norm": 1.9921875, "learning_rate": 1.700090334236676e-05, "loss": 1.2695, "step": 290 }, { "epoch": 0.2437043054427295, "grad_norm": 1.765625, "learning_rate": 1.6820234869015358e-05, "loss": 1.3047, "step": 300 }, { "epoch": 0.2518277822908205, "grad_norm": 1.8828125, "learning_rate": 1.6639566395663956e-05, "loss": 1.2727, "step": 310 }, { "epoch": 0.25995125913891143, "grad_norm": 2.140625, "learning_rate": 1.6458897922312558e-05, "loss": 1.318, "step": 320 }, { "epoch": 0.26807473598700243, "grad_norm": 1.953125, "learning_rate": 1.627822944896116e-05, "loss": 1.3477, "step": 330 }, { "epoch": 0.27619821283509344, "grad_norm": 1.921875, "learning_rate": 1.6097560975609757e-05, "loss": 1.3273, "step": 340 }, { "epoch": 0.2843216896831844, "grad_norm": 2.09375, "learning_rate": 1.5916892502258355e-05, "loss": 1.3086, "step": 350 }, { "epoch": 0.2924451665312754, "grad_norm": 1.78125, "learning_rate": 1.5736224028906957e-05, "loss": 1.2727, "step": 360 }, { "epoch": 0.3005686433793664, "grad_norm": 1.8359375, "learning_rate": 1.555555555555556e-05, "loss": 1.3195, "step": 370 }, { "epoch": 0.30869212022745735, "grad_norm": 1.8203125, "learning_rate": 1.5374887082204156e-05, "loss": 1.2828, "step": 380 }, { "epoch": 0.31681559707554835, "grad_norm": 1.7265625, "learning_rate": 1.5194218608852756e-05, "loss": 1.2719, "step": 390 }, { "epoch": 0.3249390739236393, "grad_norm": 2.03125, "learning_rate": 1.5013550135501356e-05, "loss": 1.2914, "step": 400 }, { "epoch": 0.3330625507717303, "grad_norm": 1.9140625, "learning_rate": 1.4832881662149956e-05, "loss": 1.25, "step": 410 }, { "epoch": 0.3411860276198213, "grad_norm": 1.6328125, "learning_rate": 1.4652213188798556e-05, "loss": 1.282, "step": 420 }, { "epoch": 0.34930950446791226, "grad_norm": 2.015625, "learning_rate": 1.4471544715447157e-05, "loss": 1.368, "step": 430 }, { "epoch": 0.35743298131600326, "grad_norm": 1.953125, "learning_rate": 1.4290876242095755e-05, "loss": 1.3383, "step": 440 }, { "epoch": 0.3655564581640942, "grad_norm": 1.8203125, "learning_rate": 1.4110207768744355e-05, "loss": 1.2992, "step": 450 }, { "epoch": 0.3736799350121852, "grad_norm": 1.9375, "learning_rate": 1.3929539295392955e-05, "loss": 1.3234, "step": 460 }, { "epoch": 0.3818034118602762, "grad_norm": 1.875, "learning_rate": 1.3748870822041556e-05, "loss": 1.2969, "step": 470 }, { "epoch": 0.38992688870836717, "grad_norm": 1.9140625, "learning_rate": 1.3568202348690154e-05, "loss": 1.2648, "step": 480 }, { "epoch": 0.3980503655564582, "grad_norm": 2.046875, "learning_rate": 1.3387533875338754e-05, "loss": 1.3109, "step": 490 }, { "epoch": 0.4061738424045491, "grad_norm": 1.953125, "learning_rate": 1.3206865401987354e-05, "loss": 1.3086, "step": 500 }, { "epoch": 0.41429731925264013, "grad_norm": 1.8046875, "learning_rate": 1.3026196928635954e-05, "loss": 1.2188, "step": 510 }, { "epoch": 0.42242079610073113, "grad_norm": 1.8984375, "learning_rate": 1.2845528455284555e-05, "loss": 1.2383, "step": 520 }, { "epoch": 0.4305442729488221, "grad_norm": 1.7421875, "learning_rate": 1.2664859981933153e-05, "loss": 1.3164, "step": 530 }, { "epoch": 0.4386677497969131, "grad_norm": 2.0, "learning_rate": 1.2484191508581753e-05, "loss": 1.3047, "step": 540 }, { "epoch": 0.44679122664500404, "grad_norm": 1.8359375, "learning_rate": 1.2303523035230353e-05, "loss": 1.3023, "step": 550 }, { "epoch": 0.45491470349309504, "grad_norm": 1.734375, "learning_rate": 1.2122854561878954e-05, "loss": 1.3242, "step": 560 }, { "epoch": 0.46303818034118605, "grad_norm": 2.171875, "learning_rate": 1.1942186088527553e-05, "loss": 1.2961, "step": 570 }, { "epoch": 0.471161657189277, "grad_norm": 2.0625, "learning_rate": 1.1761517615176152e-05, "loss": 1.325, "step": 580 }, { "epoch": 0.479285134037368, "grad_norm": 2.0625, "learning_rate": 1.1580849141824752e-05, "loss": 1.232, "step": 590 }, { "epoch": 0.487408610885459, "grad_norm": 2.15625, "learning_rate": 1.1400180668473354e-05, "loss": 1.2805, "step": 600 }, { "epoch": 0.49553208773354995, "grad_norm": 1.8046875, "learning_rate": 1.1219512195121953e-05, "loss": 1.2867, "step": 610 }, { "epoch": 0.503655564581641, "grad_norm": 1.765625, "learning_rate": 1.1038843721770552e-05, "loss": 1.3094, "step": 620 }, { "epoch": 0.511779041429732, "grad_norm": 1.8515625, "learning_rate": 1.0858175248419151e-05, "loss": 1.3133, "step": 630 }, { "epoch": 0.5199025182778229, "grad_norm": 1.8203125, "learning_rate": 1.0677506775067751e-05, "loss": 1.3578, "step": 640 }, { "epoch": 0.5280259951259139, "grad_norm": 1.8828125, "learning_rate": 1.0496838301716353e-05, "loss": 1.2969, "step": 650 }, { "epoch": 0.5361494719740049, "grad_norm": 1.8359375, "learning_rate": 1.031616982836495e-05, "loss": 1.2867, "step": 660 }, { "epoch": 0.5442729488220959, "grad_norm": 1.8984375, "learning_rate": 1.013550135501355e-05, "loss": 1.2977, "step": 670 }, { "epoch": 0.5523964256701869, "grad_norm": 2.0625, "learning_rate": 9.95483288166215e-06, "loss": 1.2828, "step": 680 }, { "epoch": 0.5605199025182778, "grad_norm": 1.796875, "learning_rate": 9.77416440831075e-06, "loss": 1.2984, "step": 690 }, { "epoch": 0.5686433793663688, "grad_norm": 1.796875, "learning_rate": 9.59349593495935e-06, "loss": 1.2656, "step": 700 }, { "epoch": 0.5767668562144598, "grad_norm": 1.703125, "learning_rate": 9.412827461607951e-06, "loss": 1.2453, "step": 710 }, { "epoch": 0.5848903330625508, "grad_norm": 1.859375, "learning_rate": 9.23215898825655e-06, "loss": 1.3477, "step": 720 }, { "epoch": 0.5930138099106418, "grad_norm": 2.171875, "learning_rate": 9.051490514905151e-06, "loss": 1.3141, "step": 730 }, { "epoch": 0.6011372867587328, "grad_norm": 1.84375, "learning_rate": 8.870822041553749e-06, "loss": 1.2219, "step": 740 }, { "epoch": 0.6092607636068237, "grad_norm": 1.796875, "learning_rate": 8.690153568202349e-06, "loss": 1.2766, "step": 750 }, { "epoch": 0.6173842404549147, "grad_norm": 2.0625, "learning_rate": 8.509485094850949e-06, "loss": 1.2789, "step": 760 }, { "epoch": 0.6255077173030057, "grad_norm": 2.109375, "learning_rate": 8.328816621499549e-06, "loss": 1.2828, "step": 770 }, { "epoch": 0.6336311941510967, "grad_norm": 1.9140625, "learning_rate": 8.148148148148148e-06, "loss": 1.2492, "step": 780 }, { "epoch": 0.6417546709991877, "grad_norm": 1.84375, "learning_rate": 7.967479674796748e-06, "loss": 1.2836, "step": 790 }, { "epoch": 0.6498781478472786, "grad_norm": 2.15625, "learning_rate": 7.78681120144535e-06, "loss": 1.282, "step": 800 }, { "epoch": 0.6580016246953696, "grad_norm": 1.984375, "learning_rate": 7.6061427280939486e-06, "loss": 1.268, "step": 810 }, { "epoch": 0.6661251015434606, "grad_norm": 1.8046875, "learning_rate": 7.425474254742548e-06, "loss": 1.3234, "step": 820 }, { "epoch": 0.6742485783915516, "grad_norm": 2.3125, "learning_rate": 7.244805781391147e-06, "loss": 1.3195, "step": 830 }, { "epoch": 0.6823720552396426, "grad_norm": 1.9453125, "learning_rate": 7.064137308039748e-06, "loss": 1.2617, "step": 840 }, { "epoch": 0.6904955320877335, "grad_norm": 1.84375, "learning_rate": 6.883468834688347e-06, "loss": 1.3188, "step": 850 }, { "epoch": 0.6986190089358245, "grad_norm": 1.96875, "learning_rate": 6.7028003613369475e-06, "loss": 1.2969, "step": 860 }, { "epoch": 0.7067424857839155, "grad_norm": 2.28125, "learning_rate": 6.5221318879855465e-06, "loss": 1.3078, "step": 870 }, { "epoch": 0.7148659626320065, "grad_norm": 2.0, "learning_rate": 6.341463414634147e-06, "loss": 1.2789, "step": 880 }, { "epoch": 0.7229894394800975, "grad_norm": 2.046875, "learning_rate": 6.160794941282746e-06, "loss": 1.2898, "step": 890 }, { "epoch": 0.7311129163281884, "grad_norm": 1.875, "learning_rate": 5.980126467931347e-06, "loss": 1.2625, "step": 900 }, { "epoch": 0.7392363931762794, "grad_norm": 2.0625, "learning_rate": 5.7994579945799465e-06, "loss": 1.3141, "step": 910 }, { "epoch": 0.7473598700243704, "grad_norm": 1.640625, "learning_rate": 5.618789521228546e-06, "loss": 1.2594, "step": 920 }, { "epoch": 0.7554833468724614, "grad_norm": 1.859375, "learning_rate": 5.438121047877146e-06, "loss": 1.3594, "step": 930 }, { "epoch": 0.7636068237205524, "grad_norm": 2.015625, "learning_rate": 5.257452574525745e-06, "loss": 1.2742, "step": 940 }, { "epoch": 0.7717303005686433, "grad_norm": 2.125, "learning_rate": 5.076784101174346e-06, "loss": 1.2391, "step": 950 }, { "epoch": 0.7798537774167343, "grad_norm": 1.90625, "learning_rate": 4.8961156278229455e-06, "loss": 1.3109, "step": 960 }, { "epoch": 0.7879772542648253, "grad_norm": 1.734375, "learning_rate": 4.715447154471545e-06, "loss": 1.3289, "step": 970 }, { "epoch": 0.7961007311129163, "grad_norm": 2.125, "learning_rate": 4.534778681120145e-06, "loss": 1.2641, "step": 980 }, { "epoch": 0.8042242079610074, "grad_norm": 1.8828125, "learning_rate": 4.354110207768745e-06, "loss": 1.2391, "step": 990 }, { "epoch": 0.8123476848090982, "grad_norm": 1.6875, "learning_rate": 4.173441734417345e-06, "loss": 1.2977, "step": 1000 }, { "epoch": 0.8204711616571893, "grad_norm": 2.078125, "learning_rate": 3.9927732610659445e-06, "loss": 1.2625, "step": 1010 }, { "epoch": 0.8285946385052803, "grad_norm": 1.6484375, "learning_rate": 3.812104787714544e-06, "loss": 1.2937, "step": 1020 }, { "epoch": 0.8367181153533713, "grad_norm": 1.96875, "learning_rate": 3.6314363143631437e-06, "loss": 1.3109, "step": 1030 }, { "epoch": 0.8448415922014623, "grad_norm": 1.84375, "learning_rate": 3.450767841011744e-06, "loss": 1.2125, "step": 1040 }, { "epoch": 0.8529650690495532, "grad_norm": 1.8203125, "learning_rate": 3.2700993676603437e-06, "loss": 1.2547, "step": 1050 }, { "epoch": 0.8610885458976442, "grad_norm": 1.8515625, "learning_rate": 3.0894308943089435e-06, "loss": 1.2547, "step": 1060 }, { "epoch": 0.8692120227457352, "grad_norm": 1.8984375, "learning_rate": 2.9087624209575433e-06, "loss": 1.2523, "step": 1070 }, { "epoch": 0.8773354995938262, "grad_norm": 1.8359375, "learning_rate": 2.728093947606143e-06, "loss": 1.3062, "step": 1080 }, { "epoch": 0.8854589764419172, "grad_norm": 1.7890625, "learning_rate": 2.547425474254743e-06, "loss": 1.2961, "step": 1090 }, { "epoch": 0.8935824532900081, "grad_norm": 1.7890625, "learning_rate": 2.3667570009033427e-06, "loss": 1.2406, "step": 1100 }, { "epoch": 0.9017059301380991, "grad_norm": 1.765625, "learning_rate": 2.186088527551942e-06, "loss": 1.3203, "step": 1110 }, { "epoch": 0.9098294069861901, "grad_norm": 2.03125, "learning_rate": 2.0054200542005423e-06, "loss": 1.3039, "step": 1120 }, { "epoch": 0.9179528838342811, "grad_norm": 2.09375, "learning_rate": 1.824751580849142e-06, "loss": 1.2844, "step": 1130 }, { "epoch": 0.9260763606823721, "grad_norm": 1.8359375, "learning_rate": 1.6440831074977418e-06, "loss": 1.3031, "step": 1140 }, { "epoch": 0.934199837530463, "grad_norm": 1.9609375, "learning_rate": 1.4634146341463414e-06, "loss": 1.2391, "step": 1150 }, { "epoch": 0.942323314378554, "grad_norm": 1.875, "learning_rate": 1.2827461607949414e-06, "loss": 1.2508, "step": 1160 }, { "epoch": 0.950446791226645, "grad_norm": 1.859375, "learning_rate": 1.102077687443541e-06, "loss": 1.3148, "step": 1170 }, { "epoch": 0.958570268074736, "grad_norm": 1.875, "learning_rate": 9.21409214092141e-07, "loss": 1.2883, "step": 1180 }, { "epoch": 0.966693744922827, "grad_norm": 2.03125, "learning_rate": 7.407407407407407e-07, "loss": 1.282, "step": 1190 }, { "epoch": 0.974817221770918, "grad_norm": 1.75, "learning_rate": 5.600722673893405e-07, "loss": 1.318, "step": 1200 }, { "epoch": 0.9829406986190089, "grad_norm": 2.0, "learning_rate": 3.794037940379404e-07, "loss": 1.2641, "step": 1210 }, { "epoch": 0.9910641754670999, "grad_norm": 2.1875, "learning_rate": 1.987353206865402e-07, "loss": 1.3148, "step": 1220 }, { "epoch": 0.9991876523151909, "grad_norm": 1.953125, "learning_rate": 1.806684733514002e-08, "loss": 1.2688, "step": 1230 }, { "epoch": 1.0, "step": 1231, "total_flos": 5.907000409747046e+17, "train_loss": 1.310145461007311, "train_runtime": 4855.0115, "train_samples_per_second": 8.111, "train_steps_per_second": 0.254 } ], "logging_steps": 10, "max_steps": 1231, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.907000409747046e+17, "train_batch_size": 32, "trial_name": null, "trial_params": null }