|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.30143180105501133, |
|
"eval_steps": 500, |
|
"global_step": 50, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.006028636021100226, |
|
"grad_norm": 0.486471027135849, |
|
"learning_rate": 0.0001, |
|
"loss": 0.3471, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.012057272042200452, |
|
"grad_norm": 0.5119717121124268, |
|
"learning_rate": 0.0002, |
|
"loss": 0.3925, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01808590806330068, |
|
"grad_norm": 0.4457521140575409, |
|
"learning_rate": 0.00019978589232386035, |
|
"loss": 0.3398, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.024114544084400905, |
|
"grad_norm": 0.409963995218277, |
|
"learning_rate": 0.00019914448613738106, |
|
"loss": 0.2609, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.03014318010550113, |
|
"grad_norm": 0.5352571606636047, |
|
"learning_rate": 0.00019807852804032305, |
|
"loss": 0.2358, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.03617181612660136, |
|
"grad_norm": 0.495702862739563, |
|
"learning_rate": 0.00019659258262890683, |
|
"loss": 0.3512, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.04220045214770158, |
|
"grad_norm": 0.4354080855846405, |
|
"learning_rate": 0.0001946930129495106, |
|
"loss": 0.2329, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.04822908816880181, |
|
"grad_norm": 0.3071988523006439, |
|
"learning_rate": 0.0001923879532511287, |
|
"loss": 0.1409, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.05425772418990203, |
|
"grad_norm": 0.25767603516578674, |
|
"learning_rate": 0.00018968727415326884, |
|
"loss": 0.1745, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.06028636021100226, |
|
"grad_norm": 0.24111005663871765, |
|
"learning_rate": 0.00018660254037844388, |
|
"loss": 0.1417, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06631499623210249, |
|
"grad_norm": 0.3556184470653534, |
|
"learning_rate": 0.00018314696123025454, |
|
"loss": 0.1783, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.07234363225320271, |
|
"grad_norm": 0.2842593789100647, |
|
"learning_rate": 0.00017933533402912354, |
|
"loss": 0.1272, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.07837226827430294, |
|
"grad_norm": 0.2948729693889618, |
|
"learning_rate": 0.00017518398074789775, |
|
"loss": 0.1524, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.08440090429540316, |
|
"grad_norm": 0.2625027298927307, |
|
"learning_rate": 0.00017071067811865476, |
|
"loss": 0.2062, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0904295403165034, |
|
"grad_norm": 0.2535743713378906, |
|
"learning_rate": 0.00016593458151000688, |
|
"loss": 0.1715, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.09645817633760362, |
|
"grad_norm": 0.22839832305908203, |
|
"learning_rate": 0.00016087614290087208, |
|
"loss": 0.1143, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.10248681235870384, |
|
"grad_norm": 0.23887796700000763, |
|
"learning_rate": 0.00015555702330196023, |
|
"loss": 0.1447, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.10851544837980406, |
|
"grad_norm": 0.2899695932865143, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 0.2285, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.1145440844009043, |
|
"grad_norm": 0.2376154214143753, |
|
"learning_rate": 0.00014422886902190014, |
|
"loss": 0.137, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.12057272042200452, |
|
"grad_norm": 0.35903647541999817, |
|
"learning_rate": 0.000138268343236509, |
|
"loss": 0.1455, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.12660135644310475, |
|
"grad_norm": 0.29238924384117126, |
|
"learning_rate": 0.00013214394653031616, |
|
"loss": 0.1189, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.13262999246420498, |
|
"grad_norm": 0.23753075301647186, |
|
"learning_rate": 0.00012588190451025207, |
|
"loss": 0.1442, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.1386586284853052, |
|
"grad_norm": 0.24815434217453003, |
|
"learning_rate": 0.00011950903220161285, |
|
"loss": 0.1317, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.14468726450640543, |
|
"grad_norm": 0.2568942606449127, |
|
"learning_rate": 0.00011305261922200519, |
|
"loss": 0.1653, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.15071590052750566, |
|
"grad_norm": 0.20504365861415863, |
|
"learning_rate": 0.00010654031292301432, |
|
"loss": 0.1585, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.15674453654860587, |
|
"grad_norm": 0.2357844114303589, |
|
"learning_rate": 0.0001, |
|
"loss": 0.1181, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.1627731725697061, |
|
"grad_norm": 0.32542815804481506, |
|
"learning_rate": 9.345968707698569e-05, |
|
"loss": 0.2239, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.16880180859080632, |
|
"grad_norm": 0.27462485432624817, |
|
"learning_rate": 8.694738077799488e-05, |
|
"loss": 0.1411, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.17483044461190655, |
|
"grad_norm": 0.3306806683540344, |
|
"learning_rate": 8.049096779838719e-05, |
|
"loss": 0.1945, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.1808590806330068, |
|
"grad_norm": 0.20712682604789734, |
|
"learning_rate": 7.411809548974792e-05, |
|
"loss": 0.108, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.186887716654107, |
|
"grad_norm": 0.25769761204719543, |
|
"learning_rate": 6.785605346968386e-05, |
|
"loss": 0.1467, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.19291635267520724, |
|
"grad_norm": 0.22794191539287567, |
|
"learning_rate": 6.173165676349103e-05, |
|
"loss": 0.1393, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.19894498869630747, |
|
"grad_norm": 0.22670631110668182, |
|
"learning_rate": 5.577113097809989e-05, |
|
"loss": 0.1157, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.20497362471740768, |
|
"grad_norm": 0.21888034045696259, |
|
"learning_rate": 5.000000000000002e-05, |
|
"loss": 0.1228, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.21100226073850792, |
|
"grad_norm": 0.21215690672397614, |
|
"learning_rate": 4.444297669803981e-05, |
|
"loss": 0.1263, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.21703089675960813, |
|
"grad_norm": 0.2192404419183731, |
|
"learning_rate": 3.9123857099127936e-05, |
|
"loss": 0.133, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.22305953278070836, |
|
"grad_norm": 0.25261539220809937, |
|
"learning_rate": 3.406541848999312e-05, |
|
"loss": 0.1651, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.2290881688018086, |
|
"grad_norm": 0.2244417369365692, |
|
"learning_rate": 2.9289321881345254e-05, |
|
"loss": 0.1253, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.2351168048229088, |
|
"grad_norm": 0.22674843668937683, |
|
"learning_rate": 2.4816019252102273e-05, |
|
"loss": 0.1354, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.24114544084400905, |
|
"grad_norm": 0.1793588250875473, |
|
"learning_rate": 2.0664665970876496e-05, |
|
"loss": 0.0989, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.24717407686510928, |
|
"grad_norm": 0.20526793599128723, |
|
"learning_rate": 1.6853038769745467e-05, |
|
"loss": 0.1263, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.2532027128862095, |
|
"grad_norm": 0.2529662847518921, |
|
"learning_rate": 1.339745962155613e-05, |
|
"loss": 0.1717, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.2592313489073097, |
|
"grad_norm": 0.2361597716808319, |
|
"learning_rate": 1.0312725846731175e-05, |
|
"loss": 0.1353, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.26525998492840996, |
|
"grad_norm": 0.25286707282066345, |
|
"learning_rate": 7.612046748871327e-06, |
|
"loss": 0.1714, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.2712886209495102, |
|
"grad_norm": 0.1881972849369049, |
|
"learning_rate": 5.306987050489442e-06, |
|
"loss": 0.0928, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.2773172569706104, |
|
"grad_norm": 0.16801650822162628, |
|
"learning_rate": 3.40741737109318e-06, |
|
"loss": 0.0906, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.28334589299171065, |
|
"grad_norm": 0.2937343120574951, |
|
"learning_rate": 1.921471959676957e-06, |
|
"loss": 0.2129, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.28937452901281085, |
|
"grad_norm": 0.18096446990966797, |
|
"learning_rate": 8.555138626189618e-07, |
|
"loss": 0.1005, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.29540316503391106, |
|
"grad_norm": 0.2875116467475891, |
|
"learning_rate": 2.141076761396521e-07, |
|
"loss": 0.1927, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.30143180105501133, |
|
"grad_norm": 0.21270711719989777, |
|
"learning_rate": 0.0, |
|
"loss": 0.136, |
|
"step": 50 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 50, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.6797609962530406e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|