llm-jp-3-3.7b-o1-v0.1 / trainer_state.json
Kendamarron's picture
Upload folder using huggingface_hub
5803481 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.9972714870395634,
"eval_steps": 500,
"global_step": 366,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.054570259208731244,
"grad_norm": 2.4109368924528636,
"learning_rate": 2.702702702702703e-06,
"loss": 0.7385,
"step": 10
},
{
"epoch": 0.10914051841746249,
"grad_norm": 1.411542387667167,
"learning_rate": 5.405405405405406e-06,
"loss": 0.6258,
"step": 20
},
{
"epoch": 0.16371077762619374,
"grad_norm": 1.1090603733773723,
"learning_rate": 8.108108108108109e-06,
"loss": 0.5777,
"step": 30
},
{
"epoch": 0.21828103683492497,
"grad_norm": 1.0114013426864816,
"learning_rate": 9.997948550797227e-06,
"loss": 0.5768,
"step": 40
},
{
"epoch": 0.2728512960436562,
"grad_norm": 1.0925544693817912,
"learning_rate": 9.961525153583327e-06,
"loss": 0.5571,
"step": 50
},
{
"epoch": 0.3274215552523875,
"grad_norm": 1.1229468005660994,
"learning_rate": 9.879896064123961e-06,
"loss": 0.5579,
"step": 60
},
{
"epoch": 0.3819918144611187,
"grad_norm": 1.1457306310920201,
"learning_rate": 9.75380502539778e-06,
"loss": 0.5437,
"step": 70
},
{
"epoch": 0.43656207366984995,
"grad_norm": 1.0550317300389418,
"learning_rate": 9.584400884284546e-06,
"loss": 0.5459,
"step": 80
},
{
"epoch": 0.49113233287858116,
"grad_norm": 1.1312778139498736,
"learning_rate": 9.373227124134888e-06,
"loss": 0.5455,
"step": 90
},
{
"epoch": 0.5457025920873124,
"grad_norm": 1.0356499530726737,
"learning_rate": 9.122207801708802e-06,
"loss": 0.5388,
"step": 100
},
{
"epoch": 0.6002728512960437,
"grad_norm": 1.0985324135142938,
"learning_rate": 8.833630016614976e-06,
"loss": 0.5489,
"step": 110
},
{
"epoch": 0.654843110504775,
"grad_norm": 1.1083844896783392,
"learning_rate": 8.51012307297624e-06,
"loss": 0.5442,
"step": 120
},
{
"epoch": 0.7094133697135061,
"grad_norm": 1.0400224538030558,
"learning_rate": 8.154634523184389e-06,
"loss": 0.5359,
"step": 130
},
{
"epoch": 0.7639836289222374,
"grad_norm": 1.0744202310569617,
"learning_rate": 7.77040331201572e-06,
"loss": 0.5318,
"step": 140
},
{
"epoch": 0.8185538881309686,
"grad_norm": 1.0968285492605718,
"learning_rate": 7.360930265797934e-06,
"loss": 0.5323,
"step": 150
},
{
"epoch": 0.8731241473396999,
"grad_norm": 1.021879688858344,
"learning_rate": 6.929946195508933e-06,
"loss": 0.5171,
"step": 160
},
{
"epoch": 0.927694406548431,
"grad_norm": 1.0678204884195794,
"learning_rate": 6.481377904428171e-06,
"loss": 0.5234,
"step": 170
},
{
"epoch": 0.9822646657571623,
"grad_norm": 0.9343628148722539,
"learning_rate": 6.019312410053286e-06,
"loss": 0.5209,
"step": 180
},
{
"epoch": 1.0368349249658937,
"grad_norm": 1.2427110143315105,
"learning_rate": 5.547959706265068e-06,
"loss": 0.4878,
"step": 190
},
{
"epoch": 1.0914051841746248,
"grad_norm": 1.0091906053534088,
"learning_rate": 5.071614405023938e-06,
"loss": 0.4144,
"step": 200
},
{
"epoch": 1.145975443383356,
"grad_norm": 1.007806012878209,
"learning_rate": 4.594616607090028e-06,
"loss": 0.4102,
"step": 210
},
{
"epoch": 1.2005457025920874,
"grad_norm": 0.9761625285978685,
"learning_rate": 4.121312358283464e-06,
"loss": 0.3954,
"step": 220
},
{
"epoch": 1.2551159618008185,
"grad_norm": 0.9057196288186306,
"learning_rate": 3.656014051577713e-06,
"loss": 0.4026,
"step": 230
},
{
"epoch": 1.30968622100955,
"grad_norm": 0.9923780381040166,
"learning_rate": 3.202961135812437e-06,
"loss": 0.4061,
"step": 240
},
{
"epoch": 1.364256480218281,
"grad_norm": 0.9366887214418035,
"learning_rate": 2.766281489018482e-06,
"loss": 0.4072,
"step": 250
},
{
"epoch": 1.4188267394270122,
"grad_norm": 0.9615122411717134,
"learning_rate": 2.3499538082923607e-06,
"loss": 0.4081,
"step": 260
},
{
"epoch": 1.4733969986357436,
"grad_norm": 0.9926176178093427,
"learning_rate": 1.9577713588953797e-06,
"loss": 0.4033,
"step": 270
},
{
"epoch": 1.5279672578444747,
"grad_norm": 0.940790518539519,
"learning_rate": 1.5933074128684333e-06,
"loss": 0.4045,
"step": 280
},
{
"epoch": 1.5825375170532059,
"grad_norm": 0.9731821866758915,
"learning_rate": 1.2598826920598773e-06,
"loss": 0.4101,
"step": 290
},
{
"epoch": 1.6371077762619373,
"grad_norm": 0.9806874432871104,
"learning_rate": 9.605351122011308e-07,
"loss": 0.4028,
"step": 300
},
{
"epoch": 1.6916780354706686,
"grad_norm": 0.9218297203583149,
"learning_rate": 6.979921036993042e-07,
"loss": 0.4015,
"step": 310
},
{
"epoch": 1.7462482946793996,
"grad_norm": 0.9247290967153217,
"learning_rate": 4.7464576133899043e-07,
"loss": 0.3958,
"step": 320
},
{
"epoch": 1.800818553888131,
"grad_norm": 0.9418425606930034,
"learning_rate": 2.925310493105099e-07,
"loss": 0.4015,
"step": 330
},
{
"epoch": 1.8553888130968623,
"grad_norm": 0.9321370632360508,
"learning_rate": 1.5330726014397668e-07,
"loss": 0.3974,
"step": 340
},
{
"epoch": 1.9099590723055935,
"grad_norm": 0.9562537189946939,
"learning_rate": 5.824289648152126e-08,
"loss": 0.3824,
"step": 350
},
{
"epoch": 1.9645293315143246,
"grad_norm": 0.9275908894260719,
"learning_rate": 8.204113433559202e-09,
"loss": 0.4059,
"step": 360
},
{
"epoch": 1.9972714870395634,
"step": 366,
"total_flos": 53584849797120.0,
"train_loss": 0.48227007988372134,
"train_runtime": 1900.9535,
"train_samples_per_second": 24.658,
"train_steps_per_second": 0.193
}
],
"logging_steps": 10,
"max_steps": 366,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 53584849797120.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}