dada22231's picture
Training in progress, step 42, checkpoint
914c075 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.1422121896162527,
"eval_steps": 25,
"global_step": 42,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.07223476297968397,
"grad_norm": 0.008988643996417522,
"learning_rate": 5e-05,
"loss": 11.9328,
"step": 1
},
{
"epoch": 0.07223476297968397,
"eval_loss": NaN,
"eval_runtime": 0.1775,
"eval_samples_per_second": 281.628,
"eval_steps_per_second": 73.223,
"step": 1
},
{
"epoch": 0.14446952595936793,
"grad_norm": 0.009210865013301373,
"learning_rate": 0.0001,
"loss": 11.933,
"step": 2
},
{
"epoch": 0.21670428893905191,
"grad_norm": 0.009502468630671501,
"learning_rate": 9.986128001799077e-05,
"loss": 11.9332,
"step": 3
},
{
"epoch": 0.28893905191873587,
"grad_norm": 0.00922730565071106,
"learning_rate": 9.94459753267812e-05,
"loss": 11.9325,
"step": 4
},
{
"epoch": 0.3611738148984199,
"grad_norm": 0.009089686907827854,
"learning_rate": 9.875664641789545e-05,
"loss": 11.9326,
"step": 5
},
{
"epoch": 0.43340857787810383,
"grad_norm": 0.009873133152723312,
"learning_rate": 9.779754323328192e-05,
"loss": 11.933,
"step": 6
},
{
"epoch": 0.5056433408577878,
"grad_norm": 0.011123728938400745,
"learning_rate": 9.657457896300791e-05,
"loss": 11.9327,
"step": 7
},
{
"epoch": 0.5778781038374717,
"grad_norm": 0.012063764967024326,
"learning_rate": 9.509529358847655e-05,
"loss": 11.9327,
"step": 8
},
{
"epoch": 0.6501128668171557,
"grad_norm": 0.011134549975395203,
"learning_rate": 9.336880739593416e-05,
"loss": 11.933,
"step": 9
},
{
"epoch": 0.7223476297968398,
"grad_norm": 0.01320857834070921,
"learning_rate": 9.140576474687264e-05,
"loss": 11.9324,
"step": 10
},
{
"epoch": 0.7945823927765236,
"grad_norm": 0.012652316130697727,
"learning_rate": 8.921826845200139e-05,
"loss": 11.9326,
"step": 11
},
{
"epoch": 0.8668171557562077,
"grad_norm": 0.013287683948874474,
"learning_rate": 8.681980515339464e-05,
"loss": 11.9325,
"step": 12
},
{
"epoch": 0.9390519187358917,
"grad_norm": 0.014345780946314335,
"learning_rate": 8.422516217485826e-05,
"loss": 11.9323,
"step": 13
},
{
"epoch": 1.0474040632054176,
"grad_norm": 0.02747885324060917,
"learning_rate": 8.14503363531613e-05,
"loss": 20.5712,
"step": 14
},
{
"epoch": 1.1196388261851016,
"grad_norm": 0.015461836010217667,
"learning_rate": 7.85124354122177e-05,
"loss": 11.7716,
"step": 15
},
{
"epoch": 1.1918735891647856,
"grad_norm": 0.018558837473392487,
"learning_rate": 7.542957248827961e-05,
"loss": 13.575,
"step": 16
},
{
"epoch": 1.2641083521444696,
"grad_norm": 0.015563758090138435,
"learning_rate": 7.222075445642904e-05,
"loss": 9.4655,
"step": 17
},
{
"epoch": 1.3363431151241536,
"grad_norm": 0.01990383490920067,
"learning_rate": 6.890576474687263e-05,
"loss": 12.6392,
"step": 18
},
{
"epoch": 1.4085778781038374,
"grad_norm": 0.021893231198191643,
"learning_rate": 6.550504137351576e-05,
"loss": 13.4199,
"step": 19
},
{
"epoch": 1.4808126410835214,
"grad_norm": 0.011745378375053406,
"learning_rate": 6.203955092681039e-05,
"loss": 6.7549,
"step": 20
},
{
"epoch": 1.5530474040632054,
"grad_norm": 0.026938999071717262,
"learning_rate": 5.8530659307753036e-05,
"loss": 15.6081,
"step": 21
},
{
"epoch": 1.6252821670428894,
"grad_norm": 0.01988409459590912,
"learning_rate": 5.500000000000001e-05,
"loss": 11.9145,
"step": 22
},
{
"epoch": 1.6975169300225734,
"grad_norm": 0.009685815311968327,
"learning_rate": 5.1469340692246995e-05,
"loss": 6.0489,
"step": 23
},
{
"epoch": 1.7697516930022572,
"grad_norm": 0.030963227152824402,
"learning_rate": 4.7960449073189606e-05,
"loss": 17.7904,
"step": 24
},
{
"epoch": 1.8419864559819414,
"grad_norm": 0.022139111533761024,
"learning_rate": 4.4494958626484276e-05,
"loss": 12.1609,
"step": 25
},
{
"epoch": 1.8419864559819414,
"eval_loss": NaN,
"eval_runtime": 0.1732,
"eval_samples_per_second": 288.718,
"eval_steps_per_second": 75.067,
"step": 25
},
{
"epoch": 1.9142212189616252,
"grad_norm": 0.030451903119683266,
"learning_rate": 4.109423525312738e-05,
"loss": 15.3316,
"step": 26
},
{
"epoch": 2.0225733634311513,
"grad_norm": 0.03348216786980629,
"learning_rate": 3.777924554357096e-05,
"loss": 16.8198,
"step": 27
},
{
"epoch": 2.094808126410835,
"grad_norm": 0.02313702367246151,
"learning_rate": 3.45704275117204e-05,
"loss": 11.3532,
"step": 28
},
{
"epoch": 2.1670428893905194,
"grad_norm": 0.02855922095477581,
"learning_rate": 3.1487564587782306e-05,
"loss": 13.9424,
"step": 29
},
{
"epoch": 2.239277652370203,
"grad_norm": 0.013586718589067459,
"learning_rate": 2.854966364683872e-05,
"loss": 7.0922,
"step": 30
},
{
"epoch": 2.311512415349887,
"grad_norm": 0.03476588800549507,
"learning_rate": 2.577483782514174e-05,
"loss": 14.6245,
"step": 31
},
{
"epoch": 2.383747178329571,
"grad_norm": 0.0276983343064785,
"learning_rate": 2.3180194846605367e-05,
"loss": 12.0846,
"step": 32
},
{
"epoch": 2.455981941309255,
"grad_norm": 0.017914237454533577,
"learning_rate": 2.0781731547998614e-05,
"loss": 7.768,
"step": 33
},
{
"epoch": 2.528216704288939,
"grad_norm": 0.03817462921142578,
"learning_rate": 1.8594235253127375e-05,
"loss": 16.043,
"step": 34
},
{
"epoch": 2.600451467268623,
"grad_norm": 0.02697724476456642,
"learning_rate": 1.6631192604065855e-05,
"loss": 11.9938,
"step": 35
},
{
"epoch": 2.672686230248307,
"grad_norm": 0.01927335001528263,
"learning_rate": 1.490470641152345e-05,
"loss": 9.3809,
"step": 36
},
{
"epoch": 2.744920993227991,
"grad_norm": 0.03292158618569374,
"learning_rate": 1.3425421036992098e-05,
"loss": 14.5154,
"step": 37
},
{
"epoch": 2.8171557562076748,
"grad_norm": 0.03068559244275093,
"learning_rate": 1.2202456766718093e-05,
"loss": 12.0212,
"step": 38
},
{
"epoch": 2.889390519187359,
"grad_norm": 0.03927991911768913,
"learning_rate": 1.1243353582104556e-05,
"loss": 16.2666,
"step": 39
},
{
"epoch": 2.961625282167043,
"grad_norm": 0.032981183379888535,
"learning_rate": 1.0554024673218807e-05,
"loss": 14.3015,
"step": 40
},
{
"epoch": 3.069977426636569,
"grad_norm": 0.02991604059934616,
"learning_rate": 1.0138719982009242e-05,
"loss": 12.4303,
"step": 41
},
{
"epoch": 3.1422121896162527,
"grad_norm": 0.026301313191652298,
"learning_rate": 1e-05,
"loss": 11.9028,
"step": 42
}
],
"logging_steps": 1,
"max_steps": 42,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 25,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 877486964736.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}