|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.36496350364963503, |
|
"eval_steps": 25, |
|
"global_step": 75, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.004866180048661801, |
|
"grad_norm": 0.6534222364425659, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 9.6643, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.004866180048661801, |
|
"eval_loss": 1.717971920967102, |
|
"eval_runtime": 99.4449, |
|
"eval_samples_per_second": 1.75, |
|
"eval_steps_per_second": 0.875, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.009732360097323601, |
|
"grad_norm": 0.7636905908584595, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 10.4849, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.014598540145985401, |
|
"grad_norm": 0.8352400660514832, |
|
"learning_rate": 0.0001, |
|
"loss": 10.5691, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.019464720194647202, |
|
"grad_norm": 1.1163321733474731, |
|
"learning_rate": 9.99524110790929e-05, |
|
"loss": 10.8908, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.024330900243309004, |
|
"grad_norm": 1.1510467529296875, |
|
"learning_rate": 9.980973490458728e-05, |
|
"loss": 10.8367, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.029197080291970802, |
|
"grad_norm": 1.3899632692337036, |
|
"learning_rate": 9.957224306869053e-05, |
|
"loss": 10.3491, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0340632603406326, |
|
"grad_norm": 1.2129096984863281, |
|
"learning_rate": 9.924038765061042e-05, |
|
"loss": 11.0093, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.038929440389294405, |
|
"grad_norm": 1.380240797996521, |
|
"learning_rate": 9.881480035599667e-05, |
|
"loss": 11.1512, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.043795620437956206, |
|
"grad_norm": 1.4904532432556152, |
|
"learning_rate": 9.829629131445342e-05, |
|
"loss": 11.2217, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.04866180048661801, |
|
"grad_norm": 1.634533166885376, |
|
"learning_rate": 9.768584753741134e-05, |
|
"loss": 10.5503, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0535279805352798, |
|
"grad_norm": 1.5645819902420044, |
|
"learning_rate": 9.698463103929542e-05, |
|
"loss": 10.7389, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.058394160583941604, |
|
"grad_norm": 1.600929856300354, |
|
"learning_rate": 9.619397662556435e-05, |
|
"loss": 10.5976, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.06326034063260341, |
|
"grad_norm": 1.604222059249878, |
|
"learning_rate": 9.53153893518325e-05, |
|
"loss": 10.7785, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0681265206812652, |
|
"grad_norm": 1.727508306503296, |
|
"learning_rate": 9.435054165891109e-05, |
|
"loss": 10.7356, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.072992700729927, |
|
"grad_norm": 1.7615082263946533, |
|
"learning_rate": 9.330127018922194e-05, |
|
"loss": 10.9871, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.07785888077858881, |
|
"grad_norm": 1.8271732330322266, |
|
"learning_rate": 9.21695722906443e-05, |
|
"loss": 11.0153, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0827250608272506, |
|
"grad_norm": 1.8473275899887085, |
|
"learning_rate": 9.09576022144496e-05, |
|
"loss": 11.4744, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.08759124087591241, |
|
"grad_norm": 2.002676010131836, |
|
"learning_rate": 8.966766701456177e-05, |
|
"loss": 11.1621, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.09245742092457421, |
|
"grad_norm": 2.036090850830078, |
|
"learning_rate": 8.83022221559489e-05, |
|
"loss": 11.6886, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.09732360097323602, |
|
"grad_norm": 2.1093671321868896, |
|
"learning_rate": 8.68638668405062e-05, |
|
"loss": 11.2061, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.10218978102189781, |
|
"grad_norm": 2.150874614715576, |
|
"learning_rate": 8.535533905932738e-05, |
|
"loss": 11.1087, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.1070559610705596, |
|
"grad_norm": 2.380944013595581, |
|
"learning_rate": 8.377951038078302e-05, |
|
"loss": 10.9267, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.11192214111922141, |
|
"grad_norm": 2.3676857948303223, |
|
"learning_rate": 8.213938048432697e-05, |
|
"loss": 11.7442, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.11678832116788321, |
|
"grad_norm": 2.6119561195373535, |
|
"learning_rate": 8.043807145043604e-05, |
|
"loss": 11.3183, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.12165450121654502, |
|
"grad_norm": 2.365917921066284, |
|
"learning_rate": 7.86788218175523e-05, |
|
"loss": 10.8457, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.12165450121654502, |
|
"eval_loss": 1.371411919593811, |
|
"eval_runtime": 99.9993, |
|
"eval_samples_per_second": 1.74, |
|
"eval_steps_per_second": 0.87, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.12652068126520682, |
|
"grad_norm": 2.573071241378784, |
|
"learning_rate": 7.68649804173412e-05, |
|
"loss": 10.9635, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.13138686131386862, |
|
"grad_norm": 2.5956778526306152, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 11.7278, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.1362530413625304, |
|
"grad_norm": 2.9840710163116455, |
|
"learning_rate": 7.308743066175172e-05, |
|
"loss": 11.3258, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.1411192214111922, |
|
"grad_norm": 2.692122459411621, |
|
"learning_rate": 7.113091308703498e-05, |
|
"loss": 11.1816, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.145985401459854, |
|
"grad_norm": 2.821690559387207, |
|
"learning_rate": 6.91341716182545e-05, |
|
"loss": 10.7083, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.15085158150851583, |
|
"grad_norm": 2.8995065689086914, |
|
"learning_rate": 6.710100716628344e-05, |
|
"loss": 11.7289, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.15571776155717762, |
|
"grad_norm": 2.86710524559021, |
|
"learning_rate": 6.503528997521366e-05, |
|
"loss": 11.5014, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.16058394160583941, |
|
"grad_norm": 3.1569883823394775, |
|
"learning_rate": 6.294095225512603e-05, |
|
"loss": 10.8144, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.1654501216545012, |
|
"grad_norm": 2.958052396774292, |
|
"learning_rate": 6.0821980696905146e-05, |
|
"loss": 11.5663, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.170316301703163, |
|
"grad_norm": 2.971606731414795, |
|
"learning_rate": 5.868240888334653e-05, |
|
"loss": 11.8629, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.17518248175182483, |
|
"grad_norm": 3.4037623405456543, |
|
"learning_rate": 5.6526309611002594e-05, |
|
"loss": 11.2805, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.18004866180048662, |
|
"grad_norm": 3.7285900115966797, |
|
"learning_rate": 5.435778713738292e-05, |
|
"loss": 11.6089, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.18491484184914841, |
|
"grad_norm": 3.6044297218322754, |
|
"learning_rate": 5.218096936826681e-05, |
|
"loss": 12.6734, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.1897810218978102, |
|
"grad_norm": 3.859436511993408, |
|
"learning_rate": 5e-05, |
|
"loss": 11.3451, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.19464720194647203, |
|
"grad_norm": 4.121180534362793, |
|
"learning_rate": 4.781903063173321e-05, |
|
"loss": 11.4139, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.19951338199513383, |
|
"grad_norm": 4.352361679077148, |
|
"learning_rate": 4.564221286261709e-05, |
|
"loss": 13.3278, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.20437956204379562, |
|
"grad_norm": 3.7588999271392822, |
|
"learning_rate": 4.347369038899744e-05, |
|
"loss": 10.9387, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.20924574209245742, |
|
"grad_norm": 4.562860012054443, |
|
"learning_rate": 4.131759111665349e-05, |
|
"loss": 12.128, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.2141119221411192, |
|
"grad_norm": 4.854923725128174, |
|
"learning_rate": 3.917801930309486e-05, |
|
"loss": 11.4224, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.21897810218978103, |
|
"grad_norm": 4.410684585571289, |
|
"learning_rate": 3.705904774487396e-05, |
|
"loss": 11.0198, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.22384428223844283, |
|
"grad_norm": 4.752728462219238, |
|
"learning_rate": 3.4964710024786354e-05, |
|
"loss": 11.2316, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.22871046228710462, |
|
"grad_norm": 4.889558792114258, |
|
"learning_rate": 3.289899283371657e-05, |
|
"loss": 10.6528, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.23357664233576642, |
|
"grad_norm": 6.103560924530029, |
|
"learning_rate": 3.086582838174551e-05, |
|
"loss": 11.6173, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.2384428223844282, |
|
"grad_norm": 6.686156749725342, |
|
"learning_rate": 2.886908691296504e-05, |
|
"loss": 11.7325, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.24330900243309003, |
|
"grad_norm": 8.535616874694824, |
|
"learning_rate": 2.6912569338248315e-05, |
|
"loss": 10.9532, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.24330900243309003, |
|
"eval_loss": 1.3366206884384155, |
|
"eval_runtime": 100.0065, |
|
"eval_samples_per_second": 1.74, |
|
"eval_steps_per_second": 0.87, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.24817518248175183, |
|
"grad_norm": 1.754204511642456, |
|
"learning_rate": 2.500000000000001e-05, |
|
"loss": 8.2569, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.25304136253041365, |
|
"grad_norm": 2.188532829284668, |
|
"learning_rate": 2.3135019582658802e-05, |
|
"loss": 8.7464, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.25790754257907544, |
|
"grad_norm": 2.390721559524536, |
|
"learning_rate": 2.132117818244771e-05, |
|
"loss": 9.9108, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.26277372262773724, |
|
"grad_norm": 2.576392889022827, |
|
"learning_rate": 1.9561928549563968e-05, |
|
"loss": 10.1054, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.26763990267639903, |
|
"grad_norm": 2.147494077682495, |
|
"learning_rate": 1.7860619515673033e-05, |
|
"loss": 9.8218, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.2725060827250608, |
|
"grad_norm": 2.0287535190582275, |
|
"learning_rate": 1.622048961921699e-05, |
|
"loss": 10.3552, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.2773722627737226, |
|
"grad_norm": 2.4717705249786377, |
|
"learning_rate": 1.4644660940672627e-05, |
|
"loss": 10.4594, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.2822384428223844, |
|
"grad_norm": 2.5552027225494385, |
|
"learning_rate": 1.3136133159493802e-05, |
|
"loss": 10.1801, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.2871046228710462, |
|
"grad_norm": 2.0711898803710938, |
|
"learning_rate": 1.1697777844051105e-05, |
|
"loss": 10.1532, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.291970802919708, |
|
"grad_norm": 2.061006546020508, |
|
"learning_rate": 1.0332332985438248e-05, |
|
"loss": 10.4756, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.29683698296836986, |
|
"grad_norm": 2.3004233837127686, |
|
"learning_rate": 9.042397785550405e-06, |
|
"loss": 9.6621, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.30170316301703165, |
|
"grad_norm": 2.1852986812591553, |
|
"learning_rate": 7.830427709355725e-06, |
|
"loss": 10.3212, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.30656934306569344, |
|
"grad_norm": 2.386976957321167, |
|
"learning_rate": 6.698729810778065e-06, |
|
"loss": 10.2014, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.31143552311435524, |
|
"grad_norm": 2.3682126998901367, |
|
"learning_rate": 5.649458341088915e-06, |
|
"loss": 10.4178, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.31630170316301703, |
|
"grad_norm": 2.1567893028259277, |
|
"learning_rate": 4.684610648167503e-06, |
|
"loss": 9.9836, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.32116788321167883, |
|
"grad_norm": 2.226053476333618, |
|
"learning_rate": 3.8060233744356633e-06, |
|
"loss": 10.4578, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.3260340632603406, |
|
"grad_norm": 2.296086072921753, |
|
"learning_rate": 3.0153689607045845e-06, |
|
"loss": 10.5194, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.3309002433090024, |
|
"grad_norm": 5.2668776512146, |
|
"learning_rate": 2.314152462588659e-06, |
|
"loss": 10.2628, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.3357664233576642, |
|
"grad_norm": 2.316244602203369, |
|
"learning_rate": 1.70370868554659e-06, |
|
"loss": 9.9051, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.340632603406326, |
|
"grad_norm": 2.4800219535827637, |
|
"learning_rate": 1.1851996440033319e-06, |
|
"loss": 10.9728, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.34549878345498786, |
|
"grad_norm": 2.387953281402588, |
|
"learning_rate": 7.596123493895991e-07, |
|
"loss": 10.6898, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.35036496350364965, |
|
"grad_norm": 2.5870020389556885, |
|
"learning_rate": 4.277569313094809e-07, |
|
"loss": 11.2093, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.35523114355231145, |
|
"grad_norm": 2.600900888442993, |
|
"learning_rate": 1.9026509541272275e-07, |
|
"loss": 10.8837, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.36009732360097324, |
|
"grad_norm": 2.623093366622925, |
|
"learning_rate": 4.7588920907110094e-08, |
|
"loss": 10.6526, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.36496350364963503, |
|
"grad_norm": 2.7400941848754883, |
|
"learning_rate": 0.0, |
|
"loss": 11.0227, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.36496350364963503, |
|
"eval_loss": 1.3052926063537598, |
|
"eval_runtime": 100.089, |
|
"eval_samples_per_second": 1.738, |
|
"eval_steps_per_second": 0.869, |
|
"step": 75 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 75, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.9422814666752e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|