|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 11.152748037116346, |
|
"global_step": 250000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.3999999999999997e-05, |
|
"loss": 0.8845, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.7999999999999994e-05, |
|
"loss": 0.689, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 0.6793121099472046, |
|
"eval_runtime": 2.3939, |
|
"eval_samples_per_second": 959.513, |
|
"eval_steps_per_second": 15.038, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 7.199999999999999e-05, |
|
"loss": 0.6805, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.599999999999999e-05, |
|
"loss": 0.6802, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 0.6786620616912842, |
|
"eval_runtime": 2.1637, |
|
"eval_samples_per_second": 1061.605, |
|
"eval_steps_per_second": 16.638, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00011999999999999999, |
|
"loss": 0.6799, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00014399999999999998, |
|
"loss": 0.6795, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 0.6788004040718079, |
|
"eval_runtime": 2.1379, |
|
"eval_samples_per_second": 1074.407, |
|
"eval_steps_per_second": 16.839, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.000168, |
|
"loss": 0.6791, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019199999999999998, |
|
"loss": 0.679, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 0.6781774163246155, |
|
"eval_runtime": 2.1652, |
|
"eval_samples_per_second": 1060.868, |
|
"eval_steps_per_second": 16.627, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00021599999999999996, |
|
"loss": 0.6788, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00023999999999999998, |
|
"loss": 0.6787, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 0.6782493591308594, |
|
"eval_runtime": 2.1305, |
|
"eval_samples_per_second": 1078.133, |
|
"eval_steps_per_second": 16.897, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00026399999999999997, |
|
"loss": 0.6786, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00028799999999999995, |
|
"loss": 0.6786, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_loss": 0.6780672073364258, |
|
"eval_runtime": 2.1865, |
|
"eval_samples_per_second": 1050.561, |
|
"eval_steps_per_second": 16.465, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.000312, |
|
"loss": 0.6785, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.000336, |
|
"loss": 0.6784, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 0.6781116724014282, |
|
"eval_runtime": 2.1436, |
|
"eval_samples_per_second": 1071.581, |
|
"eval_steps_per_second": 16.794, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00035999999999999997, |
|
"loss": 0.6783, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00038399999999999996, |
|
"loss": 0.6783, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 0.6780590415000916, |
|
"eval_runtime": 2.0716, |
|
"eval_samples_per_second": 1108.795, |
|
"eval_steps_per_second": 17.378, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.000408, |
|
"loss": 0.6783, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00043199999999999993, |
|
"loss": 0.6781, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 0.677251398563385, |
|
"eval_runtime": 2.0351, |
|
"eval_samples_per_second": 1128.706, |
|
"eval_steps_per_second": 17.69, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00045599999999999997, |
|
"loss": 0.6776, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00047999999999999996, |
|
"loss": 0.6775, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 0.6778020262718201, |
|
"eval_runtime": 2.2347, |
|
"eval_samples_per_second": 1027.899, |
|
"eval_steps_per_second": 16.11, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0005039999999999999, |
|
"loss": 0.6775, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0005279999999999999, |
|
"loss": 0.6775, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"eval_loss": 0.6769479513168335, |
|
"eval_runtime": 2.177, |
|
"eval_samples_per_second": 1055.099, |
|
"eval_steps_per_second": 16.536, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.000552, |
|
"loss": 0.6773, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0005759999999999999, |
|
"loss": 0.6773, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"eval_loss": 0.6773238182067871, |
|
"eval_runtime": 2.1281, |
|
"eval_samples_per_second": 1079.366, |
|
"eval_steps_per_second": 16.916, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0006, |
|
"loss": 0.6773, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.0005999935478721662, |
|
"loss": 0.6774, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"eval_loss": 0.677127480506897, |
|
"eval_runtime": 2.1773, |
|
"eval_samples_per_second": 1054.986, |
|
"eval_steps_per_second": 16.534, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.000599974191770902, |
|
"loss": 0.6773, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.0005999419325429058, |
|
"loss": 0.6773, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"eval_loss": 0.6771531105041504, |
|
"eval_runtime": 2.1173, |
|
"eval_samples_per_second": 1084.86, |
|
"eval_steps_per_second": 17.003, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.0005998967715993009, |
|
"loss": 0.6773, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.0005998387109155732, |
|
"loss": 0.6773, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"eval_loss": 0.6771678924560547, |
|
"eval_runtime": 2.1919, |
|
"eval_samples_per_second": 1047.963, |
|
"eval_steps_per_second": 16.424, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.000599767753031485, |
|
"loss": 0.6773, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.0005996839010509641, |
|
"loss": 0.6772, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"eval_loss": 0.6776318550109863, |
|
"eval_runtime": 2.199, |
|
"eval_samples_per_second": 1044.559, |
|
"eval_steps_per_second": 16.371, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0005995871586419678, |
|
"loss": 0.6773, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.0005994775300363225, |
|
"loss": 0.6773, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"eval_loss": 0.676984429359436, |
|
"eval_runtime": 2.1946, |
|
"eval_samples_per_second": 1046.652, |
|
"eval_steps_per_second": 16.404, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.0005993550200295384, |
|
"loss": 0.6772, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0005992196339806002, |
|
"loss": 0.6772, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"eval_loss": 0.6774880290031433, |
|
"eval_runtime": 2.1027, |
|
"eval_samples_per_second": 1092.415, |
|
"eval_steps_per_second": 17.121, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.0005990713778117324, |
|
"loss": 0.6773, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.0005989102580081398, |
|
"loss": 0.6772, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"eval_loss": 0.676984965801239, |
|
"eval_runtime": 2.1686, |
|
"eval_samples_per_second": 1059.223, |
|
"eval_steps_per_second": 16.601, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.0005987362816177249, |
|
"loss": 0.6773, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.0005985494562507783, |
|
"loss": 0.6774, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"eval_loss": 0.6769698262214661, |
|
"eval_runtime": 2.1456, |
|
"eval_samples_per_second": 1070.586, |
|
"eval_steps_per_second": 16.779, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.000598349790079647, |
|
"loss": 0.6773, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.000598137291838376, |
|
"loss": 0.6772, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"eval_loss": 0.6761835217475891, |
|
"eval_runtime": 2.1367, |
|
"eval_samples_per_second": 1075.027, |
|
"eval_steps_per_second": 16.848, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.000597911970822327, |
|
"loss": 0.6772, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.000597673836887771, |
|
"loss": 0.6773, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"eval_loss": 0.6774830222129822, |
|
"eval_runtime": 2.1107, |
|
"eval_samples_per_second": 1088.282, |
|
"eval_steps_per_second": 17.056, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.0005974229004514577, |
|
"loss": 0.6773, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.0005971591724901598, |
|
"loss": 0.6773, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"eval_loss": 0.6764113903045654, |
|
"eval_runtime": 2.206, |
|
"eval_samples_per_second": 1041.236, |
|
"eval_steps_per_second": 16.319, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.0005968826645401927, |
|
"loss": 0.6772, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00059659338869691, |
|
"loss": 0.6772, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"eval_loss": 0.6768463253974915, |
|
"eval_runtime": 2.2173, |
|
"eval_samples_per_second": 1035.939, |
|
"eval_steps_per_second": 16.236, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.0005962913576141742, |
|
"loss": 0.6773, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.000595976584503803, |
|
"loss": 0.6772, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"eval_loss": 0.6769193410873413, |
|
"eval_runtime": 2.2077, |
|
"eval_samples_per_second": 1040.465, |
|
"eval_steps_per_second": 16.307, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.0005956490831349923, |
|
"loss": 0.6773, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.0005953088678337129, |
|
"loss": 0.6772, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"eval_loss": 0.6775402426719666, |
|
"eval_runtime": 2.2094, |
|
"eval_samples_per_second": 1039.671, |
|
"eval_steps_per_second": 16.294, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.0005949559534820841, |
|
"loss": 0.6772, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.0005945903555177229, |
|
"loss": 0.6772, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"eval_loss": 0.6775718331336975, |
|
"eval_runtime": 2.1372, |
|
"eval_samples_per_second": 1074.76, |
|
"eval_steps_per_second": 16.844, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.0005942120899330687, |
|
"loss": 0.6772, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.0005938211732746836, |
|
"loss": 0.6772, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"eval_loss": 0.6771659255027771, |
|
"eval_runtime": 2.191, |
|
"eval_samples_per_second": 1048.378, |
|
"eval_steps_per_second": 16.431, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.0005934176226425286, |
|
"loss": 0.6772, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.0005930014556892158, |
|
"loss": 0.6772, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"eval_loss": 0.6768732666969299, |
|
"eval_runtime": 2.2912, |
|
"eval_samples_per_second": 1002.553, |
|
"eval_steps_per_second": 15.713, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.0005925726906192357, |
|
"loss": 0.6772, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 0.0005921313461881617, |
|
"loss": 0.6773, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"eval_loss": 0.6772350072860718, |
|
"eval_runtime": 2.2465, |
|
"eval_samples_per_second": 1022.498, |
|
"eval_steps_per_second": 16.025, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 0.0005916774417018287, |
|
"loss": 0.6772, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 0.0005912109970154897, |
|
"loss": 0.6772, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"eval_loss": 0.6777171492576599, |
|
"eval_runtime": 2.1569, |
|
"eval_samples_per_second": 1064.935, |
|
"eval_steps_per_second": 16.69, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 0.0005907320325329461, |
|
"loss": 0.6772, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 0.0005902405692056561, |
|
"loss": 0.6772, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"eval_loss": 0.676947832107544, |
|
"eval_runtime": 2.3119, |
|
"eval_samples_per_second": 993.573, |
|
"eval_steps_per_second": 15.572, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 0.0005897366285318178, |
|
"loss": 0.6772, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 0.0005892202325554288, |
|
"loss": 0.6773, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"eval_loss": 0.6766595840454102, |
|
"eval_runtime": 2.1462, |
|
"eval_samples_per_second": 1070.248, |
|
"eval_steps_per_second": 16.774, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 0.0005886914038653217, |
|
"loss": 0.6772, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 0.0005881501655941771, |
|
"loss": 0.677, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"eval_loss": 0.6765649914741516, |
|
"eval_runtime": 2.1369, |
|
"eval_samples_per_second": 1074.929, |
|
"eval_steps_per_second": 16.847, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 0.00058759654141751, |
|
"loss": 0.6768, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 0.0005870305555526355, |
|
"loss": 0.6765, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"eval_loss": 0.676632821559906, |
|
"eval_runtime": 2.2745, |
|
"eval_samples_per_second": 1009.888, |
|
"eval_steps_per_second": 15.828, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 0.0005864522327576088, |
|
"loss": 0.6764, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 0.0005858615983301424, |
|
"loss": 0.6763, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"eval_loss": 0.6765820980072021, |
|
"eval_runtime": 2.1766, |
|
"eval_samples_per_second": 1055.304, |
|
"eval_steps_per_second": 16.539, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 0.0005852586781064997, |
|
"loss": 0.6763, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 0.0005846434984603645, |
|
"loss": 0.6764, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"eval_loss": 0.6758345365524292, |
|
"eval_runtime": 2.237, |
|
"eval_samples_per_second": 1026.821, |
|
"eval_steps_per_second": 16.093, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 0.0005840160863016872, |
|
"loss": 0.6763, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 0.0005833764690755083, |
|
"loss": 0.6764, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"eval_loss": 0.6762025952339172, |
|
"eval_runtime": 2.1722, |
|
"eval_samples_per_second": 1057.45, |
|
"eval_steps_per_second": 16.573, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 0.0005827246747607574, |
|
"loss": 0.6765, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 0.0005820607318690293, |
|
"loss": 0.6758, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"eval_loss": 0.6771443486213684, |
|
"eval_runtime": 2.2147, |
|
"eval_samples_per_second": 1037.153, |
|
"eval_steps_per_second": 16.255, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 0.0005813846694433368, |
|
"loss": 0.6773, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 0.0005806965170568409, |
|
"loss": 0.6772, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"eval_loss": 0.6770426630973816, |
|
"eval_runtime": 2.1711, |
|
"eval_samples_per_second": 1057.979, |
|
"eval_steps_per_second": 16.581, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 0.0005799963048115559, |
|
"loss": 0.6757, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 0.0005792840633370341, |
|
"loss": 0.6575, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"eval_loss": 0.6465174555778503, |
|
"eval_runtime": 2.2243, |
|
"eval_samples_per_second": 1032.694, |
|
"eval_steps_per_second": 16.185, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 0.0005785598237890247, |
|
"loss": 0.6461, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 0.0005778236178481119, |
|
"loss": 0.6373, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"eval_loss": 0.631793737411499, |
|
"eval_runtime": 2.22, |
|
"eval_samples_per_second": 1034.694, |
|
"eval_steps_per_second": 16.216, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 0.0005770754777183285, |
|
"loss": 0.6309, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 0.0005763154361257473, |
|
"loss": 0.6257, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"eval_loss": 0.6184197664260864, |
|
"eval_runtime": 2.2848, |
|
"eval_samples_per_second": 1005.32, |
|
"eval_steps_per_second": 15.756, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 0.0005755435263170498, |
|
"loss": 0.623, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 0.0005747597820580717, |
|
"loss": 0.621, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"eval_loss": 0.6136025786399841, |
|
"eval_runtime": 2.2706, |
|
"eval_samples_per_second": 1011.621, |
|
"eval_steps_per_second": 15.855, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 0.000573964237632326, |
|
"loss": 0.6195, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 0.0005731569278395029, |
|
"loss": 0.6183, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"eval_loss": 0.6127315163612366, |
|
"eval_runtime": 2.1317, |
|
"eval_samples_per_second": 1077.566, |
|
"eval_steps_per_second": 16.888, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 0.0005723378879939481, |
|
"loss": 0.6172, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 0.0005715071539231178, |
|
"loss": 0.6165, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"eval_loss": 0.6103396415710449, |
|
"eval_runtime": 2.1687, |
|
"eval_samples_per_second": 1059.145, |
|
"eval_steps_per_second": 16.6, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 0.0005706647619660116, |
|
"loss": 0.6155, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 0.0005698107489715823, |
|
"loss": 0.612, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"eval_loss": 0.6013069748878479, |
|
"eval_runtime": 2.2612, |
|
"eval_samples_per_second": 1015.821, |
|
"eval_steps_per_second": 15.921, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 0.0005689451522971252, |
|
"loss": 0.6068, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 0.0005680680098066429, |
|
"loss": 0.6037, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"eval_loss": 0.5942632555961609, |
|
"eval_runtime": 2.2238, |
|
"eval_samples_per_second": 1032.912, |
|
"eval_steps_per_second": 16.188, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 0.0005671793598691895, |
|
"loss": 0.6015, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 0.0005662792413571921, |
|
"loss": 0.6, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"eval_loss": 0.5915176868438721, |
|
"eval_runtime": 2.2389, |
|
"eval_samples_per_second": 1025.951, |
|
"eval_steps_per_second": 16.079, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 0.0005653676936447504, |
|
"loss": 0.5986, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 0.0005644447566059142, |
|
"loss": 0.5973, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"eval_loss": 0.5880685448646545, |
|
"eval_runtime": 2.2056, |
|
"eval_samples_per_second": 1041.459, |
|
"eval_steps_per_second": 16.322, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 0.0005635104706129397, |
|
"loss": 0.5958, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 0.0005625648765345228, |
|
"loss": 0.5924, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"eval_loss": 0.5799189209938049, |
|
"eval_runtime": 2.2525, |
|
"eval_samples_per_second": 1019.737, |
|
"eval_steps_per_second": 15.982, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 0.0005616080157340118, |
|
"loss": 0.5872, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 0.0005606399300675978, |
|
"loss": 0.5817, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"eval_loss": 0.5670354962348938, |
|
"eval_runtime": 2.3437, |
|
"eval_samples_per_second": 980.093, |
|
"eval_steps_per_second": 15.361, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 0.0005596606618824843, |
|
"loss": 0.5766, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 0.0005586702540150338, |
|
"loss": 0.5719, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"eval_loss": 0.5556795597076416, |
|
"eval_runtime": 2.1734, |
|
"eval_samples_per_second": 1056.846, |
|
"eval_steps_per_second": 16.564, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 0.000557668749788895, |
|
"loss": 0.5681, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 0.0005566561930131072, |
|
"loss": 0.5651, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"eval_loss": 0.5476705431938171, |
|
"eval_runtime": 2.2374, |
|
"eval_samples_per_second": 1026.619, |
|
"eval_steps_per_second": 16.09, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 0.000555632627980184, |
|
"loss": 0.562, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 0.0005545980994641758, |
|
"loss": 0.5592, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"eval_loss": 0.5407888889312744, |
|
"eval_runtime": 2.2932, |
|
"eval_samples_per_second": 1001.641, |
|
"eval_steps_per_second": 15.698, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 0.0005535526527187115, |
|
"loss": 0.5563, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 0.0005524963334750183, |
|
"loss": 0.5534, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"eval_loss": 0.536229133605957, |
|
"eval_runtime": 2.1818, |
|
"eval_samples_per_second": 1052.81, |
|
"eval_steps_per_second": 16.5, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 0.0005514291879399219, |
|
"loss": 0.5496, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 0.000550351262793825, |
|
"loss": 0.5446, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"eval_loss": 0.5251114964485168, |
|
"eval_runtime": 2.298, |
|
"eval_samples_per_second": 999.556, |
|
"eval_steps_per_second": 15.666, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 0.0005492626051886659, |
|
"loss": 0.5392, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 0.0005481632627458546, |
|
"loss": 0.5342, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"eval_loss": 0.5129896402359009, |
|
"eval_runtime": 2.2194, |
|
"eval_samples_per_second": 1034.968, |
|
"eval_steps_per_second": 16.221, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 0.0005470532835541911, |
|
"loss": 0.529, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 0.000545932716167761, |
|
"loss": 0.5239, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"eval_loss": 0.5024312734603882, |
|
"eval_runtime": 2.205, |
|
"eval_samples_per_second": 1041.743, |
|
"eval_steps_per_second": 16.327, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 0.000544801609603812, |
|
"loss": 0.5193, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 0.0005436600133406095, |
|
"loss": 0.5147, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"eval_loss": 0.49465516209602356, |
|
"eval_runtime": 2.2427, |
|
"eval_samples_per_second": 1024.23, |
|
"eval_steps_per_second": 16.052, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 0.0005425079773152721, |
|
"loss": 0.5102, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 0.0005413455519215879, |
|
"loss": 0.5061, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"eval_loss": 0.4848201274871826, |
|
"eval_runtime": 2.1727, |
|
"eval_samples_per_second": 1057.234, |
|
"eval_steps_per_second": 16.57, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 0.0005401727880078093, |
|
"loss": 0.502, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 0.0005389897368744289, |
|
"loss": 0.4981, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"eval_loss": 0.4745917320251465, |
|
"eval_runtime": 2.1553, |
|
"eval_samples_per_second": 1065.724, |
|
"eval_steps_per_second": 16.703, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 0.0005377964502719361, |
|
"loss": 0.4972, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 0.0005365929803985524, |
|
"loss": 0.4912, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"eval_loss": 0.46808210015296936, |
|
"eval_runtime": 2.2693, |
|
"eval_samples_per_second": 1012.199, |
|
"eval_steps_per_second": 15.864, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 0.0005353793798979489, |
|
"loss": 0.4882, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 0.000534155701856943, |
|
"loss": 0.4847, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"eval_loss": 0.45994389057159424, |
|
"eval_runtime": 2.2357, |
|
"eval_samples_per_second": 1027.41, |
|
"eval_steps_per_second": 16.102, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 0.0005329219998031763, |
|
"loss": 0.4818, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 0.0005316783277027734, |
|
"loss": 0.4792, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"eval_loss": 0.4537006914615631, |
|
"eval_runtime": 2.1952, |
|
"eval_samples_per_second": 1046.372, |
|
"eval_steps_per_second": 16.399, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 0.0005304247399579808, |
|
"loss": 0.4766, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 0.0005291612914047876, |
|
"loss": 0.474, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"eval_loss": 0.44910311698913574, |
|
"eval_runtime": 2.304, |
|
"eval_samples_per_second": 996.96, |
|
"eval_steps_per_second": 15.625, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 0.0005278880373105263, |
|
"loss": 0.4713, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 0.0005266050333714561, |
|
"loss": 0.4688, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"eval_loss": 0.4437292516231537, |
|
"eval_runtime": 2.2404, |
|
"eval_samples_per_second": 1025.255, |
|
"eval_steps_per_second": 16.068, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 0.0005253123357103253, |
|
"loss": 0.4665, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 0.0005240100008739177, |
|
"loss": 0.464, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"eval_loss": 0.4391787052154541, |
|
"eval_runtime": 2.1873, |
|
"eval_samples_per_second": 1050.17, |
|
"eval_steps_per_second": 16.459, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 0.0005226980858305778, |
|
"loss": 0.4616, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 0.0005213766479677197, |
|
"loss": 0.4592, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"eval_loss": 0.43239352107048035, |
|
"eval_runtime": 2.1819, |
|
"eval_samples_per_second": 1052.771, |
|
"eval_steps_per_second": 16.5, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 0.0005200457450893163, |
|
"loss": 0.457, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 0.0005187054354133712, |
|
"loss": 0.4547, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"eval_loss": 0.4284396469593048, |
|
"eval_runtime": 2.1519, |
|
"eval_samples_per_second": 1067.405, |
|
"eval_steps_per_second": 16.729, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 0.0005173557775693715, |
|
"loss": 0.4528, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 0.0005159968305957235, |
|
"loss": 0.4507, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"eval_loss": 0.42597973346710205, |
|
"eval_runtime": 2.1768, |
|
"eval_samples_per_second": 1055.242, |
|
"eval_steps_per_second": 16.538, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 0.0005146286539371703, |
|
"loss": 0.4489, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 0.0005132513074421913, |
|
"loss": 0.4468, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"eval_loss": 0.4192256033420563, |
|
"eval_runtime": 2.2165, |
|
"eval_samples_per_second": 1036.301, |
|
"eval_steps_per_second": 16.242, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 0.0005118648513603841, |
|
"loss": 0.445, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 0.0005104693463398293, |
|
"loss": 0.4432, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"eval_loss": 0.4161074459552765, |
|
"eval_runtime": 2.2735, |
|
"eval_samples_per_second": 1010.341, |
|
"eval_steps_per_second": 15.835, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 0.0005090648534244371, |
|
"loss": 0.4415, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 0.0005076514340512776, |
|
"loss": 0.44, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"eval_loss": 0.41528481245040894, |
|
"eval_runtime": 2.1341, |
|
"eval_samples_per_second": 1076.31, |
|
"eval_steps_per_second": 16.869, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 0.0005062291500478931, |
|
"loss": 0.4386, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 0.0005047980636295937, |
|
"loss": 0.4367, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"eval_loss": 0.4101610779762268, |
|
"eval_runtime": 2.2225, |
|
"eval_samples_per_second": 1033.52, |
|
"eval_steps_per_second": 16.198, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 0.0005033582373967348, |
|
"loss": 0.4351, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 0.0005019097343319809, |
|
"loss": 0.4337, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"eval_loss": 0.40619799494743347, |
|
"eval_runtime": 2.1631, |
|
"eval_samples_per_second": 1061.919, |
|
"eval_steps_per_second": 16.643, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 0.0005004526177975481, |
|
"loss": 0.4322, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 0.0004989869515324342, |
|
"loss": 0.4311, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"eval_loss": 0.4019022583961487, |
|
"eval_runtime": 2.2694, |
|
"eval_samples_per_second": 1012.161, |
|
"eval_steps_per_second": 15.863, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 0.0004975127996496297, |
|
"loss": 0.4294, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 0.0004960302266333135, |
|
"loss": 0.4286, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"eval_loss": 0.40067020058631897, |
|
"eval_runtime": 2.31, |
|
"eval_samples_per_second": 994.37, |
|
"eval_steps_per_second": 15.584, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 0.0004945392973360323, |
|
"loss": 0.427, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 0.0004930400769758634, |
|
"loss": 0.4259, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"eval_loss": 0.3996644616127014, |
|
"eval_runtime": 2.1788, |
|
"eval_samples_per_second": 1054.255, |
|
"eval_steps_per_second": 16.523, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 0.0004915326311335622, |
|
"loss": 0.425, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 0.0004900170257496933, |
|
"loss": 0.4239, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"eval_loss": 0.3968483507633209, |
|
"eval_runtime": 2.2257, |
|
"eval_samples_per_second": 1032.02, |
|
"eval_steps_per_second": 16.174, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 0.0004884933271217461, |
|
"loss": 0.4229, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 0.0004869616019012347, |
|
"loss": 0.4218, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"eval_loss": 0.3948546051979065, |
|
"eval_runtime": 2.2856, |
|
"eval_samples_per_second": 1005.007, |
|
"eval_steps_per_second": 15.751, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 0.00048542191709078226, |
|
"loss": 0.4209, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 0.00048387434004119027, |
|
"loss": 0.4201, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"eval_loss": 0.39345651865005493, |
|
"eval_runtime": 2.2871, |
|
"eval_samples_per_second": 1004.335, |
|
"eval_steps_per_second": 15.741, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 0.0004823189384484924, |
|
"loss": 0.4206, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 0.000480755780350993, |
|
"loss": 0.4182, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"eval_loss": 0.39259859919548035, |
|
"eval_runtime": 2.2778, |
|
"eval_samples_per_second": 1008.432, |
|
"eval_steps_per_second": 15.805, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 0.0004791849341262914, |
|
"loss": 0.4176, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 0.0004776064684882901, |
|
"loss": 0.4168, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"eval_loss": 0.38789093494415283, |
|
"eval_runtime": 2.2568, |
|
"eval_samples_per_second": 1017.832, |
|
"eval_steps_per_second": 15.952, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 0.0004760204524841897, |
|
"loss": 0.416, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 0.0004744269554914683, |
|
"loss": 0.4155, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"eval_loss": 0.3885124921798706, |
|
"eval_runtime": 2.2513, |
|
"eval_samples_per_second": 1020.282, |
|
"eval_steps_per_second": 15.99, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 0.00047282604721484643, |
|
"loss": 0.4144, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 0.0004712177976832385, |
|
"loss": 0.4136, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"eval_loss": 0.38439127802848816, |
|
"eval_runtime": 2.2526, |
|
"eval_samples_per_second": 1019.704, |
|
"eval_steps_per_second": 15.981, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 0.0004696022772466888, |
|
"loss": 0.413, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 0.00046797955657329487, |
|
"loss": 0.4124, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"eval_loss": 0.3855268955230713, |
|
"eval_runtime": 2.2218, |
|
"eval_samples_per_second": 1033.851, |
|
"eval_steps_per_second": 16.203, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 0.000466349706646116, |
|
"loss": 0.4118, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 0.00046471279876006763, |
|
"loss": 0.4116, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"eval_loss": 0.3830411732196808, |
|
"eval_runtime": 2.2947, |
|
"eval_samples_per_second": 1001.012, |
|
"eval_steps_per_second": 15.688, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 0.00046306890451880395, |
|
"loss": 0.4107, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 0.0004614180958315844, |
|
"loss": 0.4098, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"eval_loss": 0.38374021649360657, |
|
"eval_runtime": 2.1424, |
|
"eval_samples_per_second": 1072.163, |
|
"eval_steps_per_second": 16.804, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 0.00045976044491012884, |
|
"loss": 0.4094, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 0.00045809602426545847, |
|
"loss": 0.4087, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"eval_loss": 0.3802391290664673, |
|
"eval_runtime": 2.1802, |
|
"eval_samples_per_second": 1053.586, |
|
"eval_steps_per_second": 16.512, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 0.00045642490670472436, |
|
"loss": 0.4084, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 0.0004547471653280225, |
|
"loss": 0.4078, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"eval_loss": 0.37994059920310974, |
|
"eval_runtime": 2.2046, |
|
"eval_samples_per_second": 1041.911, |
|
"eval_steps_per_second": 16.329, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 0.00045306287352519543, |
|
"loss": 0.407, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 0.00045137210497262333, |
|
"loss": 0.4068, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"eval_loss": 0.3793868124485016, |
|
"eval_runtime": 2.2708, |
|
"eval_samples_per_second": 1011.534, |
|
"eval_steps_per_second": 15.853, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 0.0004496749336299999, |
|
"loss": 0.4063, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 0.0004479714337370977, |
|
"loss": 0.4057, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"eval_loss": 0.37835967540740967, |
|
"eval_runtime": 2.237, |
|
"eval_samples_per_second": 1026.834, |
|
"eval_steps_per_second": 16.093, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 0.00044626167981052036, |
|
"loss": 0.4052, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 0.00044454574664044404, |
|
"loss": 0.4047, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"eval_loss": 0.37884432077407837, |
|
"eval_runtime": 2.2678, |
|
"eval_samples_per_second": 1012.88, |
|
"eval_steps_per_second": 15.874, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 0.000442823709287344, |
|
"loss": 0.4044, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 0.0004410956430787129, |
|
"loss": 0.4047, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"eval_loss": 0.37695789337158203, |
|
"eval_runtime": 2.1458, |
|
"eval_samples_per_second": 1070.466, |
|
"eval_steps_per_second": 16.777, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 0.0004393616236057647, |
|
"loss": 0.4036, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 0.00043762172672012875, |
|
"loss": 0.4029, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"eval_loss": 0.37500157952308655, |
|
"eval_runtime": 2.2301, |
|
"eval_samples_per_second": 1029.999, |
|
"eval_steps_per_second": 16.143, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 0.0004358760285305312, |
|
"loss": 0.4025, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 0.0004341246053994663, |
|
"loss": 0.4022, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"eval_loss": 0.37471264600753784, |
|
"eval_runtime": 2.2149, |
|
"eval_samples_per_second": 1037.069, |
|
"eval_steps_per_second": 16.254, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 0.00043236753393985534, |
|
"loss": 0.4019, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 0.0004306048910116964, |
|
"loss": 0.4015, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"eval_loss": 0.3735538125038147, |
|
"eval_runtime": 2.2361, |
|
"eval_samples_per_second": 1027.233, |
|
"eval_steps_per_second": 16.099, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 0.0004288367537187012, |
|
"loss": 0.4011, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 0.00042706319940492284, |
|
"loss": 0.4007, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"eval_loss": 0.3751888573169708, |
|
"eval_runtime": 2.2595, |
|
"eval_samples_per_second": 1016.613, |
|
"eval_steps_per_second": 15.933, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 0.00042528430565137254, |
|
"loss": 0.4003, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 0.00042350015027262593, |
|
"loss": 0.4, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"eval_loss": 0.37434616684913635, |
|
"eval_runtime": 2.2061, |
|
"eval_samples_per_second": 1041.22, |
|
"eval_steps_per_second": 16.319, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 0.00042171081131341917, |
|
"loss": 0.3996, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 0.00041991636704523497, |
|
"loss": 0.3995, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"eval_loss": 0.37412840127944946, |
|
"eval_runtime": 2.2118, |
|
"eval_samples_per_second": 1038.514, |
|
"eval_steps_per_second": 16.276, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 0.00041811689596287893, |
|
"loss": 0.3989, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 0.0004163124767810454, |
|
"loss": 0.3985, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"eval_loss": 0.37019699811935425, |
|
"eval_runtime": 2.207, |
|
"eval_samples_per_second": 1040.78, |
|
"eval_steps_per_second": 16.312, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 0.00041450318843087506, |
|
"loss": 0.3983, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 0.00041268911005650166, |
|
"loss": 0.3981, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"eval_loss": 0.3799527883529663, |
|
"eval_runtime": 2.7041, |
|
"eval_samples_per_second": 849.448, |
|
"eval_steps_per_second": 13.313, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 0.00041087032101159006, |
|
"loss": 0.3978, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 0.00040904690085586515, |
|
"loss": 0.3986, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"eval_loss": 0.3734039068222046, |
|
"eval_runtime": 2.2569, |
|
"eval_samples_per_second": 1017.749, |
|
"eval_steps_per_second": 15.951, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 0.0004072189293516316, |
|
"loss": 0.3969, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 0.0004053864864602847, |
|
"loss": 0.3966, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"eval_loss": 0.37050846219062805, |
|
"eval_runtime": 2.2224, |
|
"eval_samples_per_second": 1033.581, |
|
"eval_steps_per_second": 16.199, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 0.00040354965233881297, |
|
"loss": 0.3961, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 0.0004017085073362913, |
|
"loss": 0.3957, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"eval_loss": 0.36802881956100464, |
|
"eval_runtime": 2.2556, |
|
"eval_samples_per_second": 1018.364, |
|
"eval_steps_per_second": 15.96, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 0.00039986313199036664, |
|
"loss": 0.3956, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 0.00039801360702373484, |
|
"loss": 0.3957, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"eval_loss": 0.3662741482257843, |
|
"eval_runtime": 2.5144, |
|
"eval_samples_per_second": 913.525, |
|
"eval_steps_per_second": 14.317, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 0.00039616001334060954, |
|
"loss": 0.3951, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 0.00039430243202318314, |
|
"loss": 0.3948, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"eval_loss": 0.36831599473953247, |
|
"eval_runtime": 2.2146, |
|
"eval_samples_per_second": 1037.206, |
|
"eval_steps_per_second": 16.256, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 0.00039244094432808034, |
|
"loss": 0.3945, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 0.0003905756316828033, |
|
"loss": 0.3943, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"eval_loss": 0.36974549293518066, |
|
"eval_runtime": 2.1759, |
|
"eval_samples_per_second": 1055.667, |
|
"eval_steps_per_second": 16.545, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 0.00038870657568216963, |
|
"loss": 0.3939, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 0.00038683385808474416, |
|
"loss": 0.3936, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"eval_loss": 0.3671797811985016, |
|
"eval_runtime": 2.2064, |
|
"eval_samples_per_second": 1041.083, |
|
"eval_steps_per_second": 16.316, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 0.00038495756080926107, |
|
"loss": 0.3933, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 0.0003830777659310416, |
|
"loss": 0.3932, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"eval_loss": 0.36492955684661865, |
|
"eval_runtime": 2.2207, |
|
"eval_samples_per_second": 1034.354, |
|
"eval_steps_per_second": 16.211, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 0.0003811945556784033, |
|
"loss": 0.3931, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 0.00037930801242906366, |
|
"loss": 0.3925, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.36507272720336914, |
|
"eval_runtime": 2.3005, |
|
"eval_samples_per_second": 998.488, |
|
"eval_steps_per_second": 15.649, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 0.00037741821870653576, |
|
"loss": 0.3924, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 0.00037552525717651905, |
|
"loss": 0.3919, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"eval_loss": 0.3649584650993347, |
|
"eval_runtime": 2.1025, |
|
"eval_samples_per_second": 1092.533, |
|
"eval_steps_per_second": 17.123, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 0.0003736292106432834, |
|
"loss": 0.3918, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 0.00037173016204604647, |
|
"loss": 0.3915, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"eval_loss": 0.36355069279670715, |
|
"eval_runtime": 2.1224, |
|
"eval_samples_per_second": 1082.26, |
|
"eval_steps_per_second": 16.962, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 0.0003698281944553456, |
|
"loss": 0.3915, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 0.0003679233910694053, |
|
"loss": 0.3911, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"eval_loss": 0.36546093225479126, |
|
"eval_runtime": 2.202, |
|
"eval_samples_per_second": 1043.148, |
|
"eval_steps_per_second": 16.349, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 5.15, |
|
"learning_rate": 0.000366015835210496, |
|
"loss": 0.3911, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 0.0003641056103212908, |
|
"loss": 0.3905, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"eval_loss": 0.36497965455055237, |
|
"eval_runtime": 2.1541, |
|
"eval_samples_per_second": 1066.328, |
|
"eval_steps_per_second": 16.712, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 0.00036219279996121446, |
|
"loss": 0.3904, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 0.0003602774878027888, |
|
"loss": 0.3905, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"eval_loss": 0.40538668632507324, |
|
"eval_runtime": 2.1607, |
|
"eval_samples_per_second": 1063.067, |
|
"eval_steps_per_second": 16.661, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 0.00035835975762797245, |
|
"loss": 0.3917, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 0.0003564396933244957, |
|
"loss": 0.3894, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"eval_loss": 0.36090952157974243, |
|
"eval_runtime": 2.2235, |
|
"eval_samples_per_second": 1033.034, |
|
"eval_steps_per_second": 16.19, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 0.0003545173788821915, |
|
"loss": 0.3893, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 0.00035259289838932104, |
|
"loss": 0.3889, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"eval_loss": 0.3599041998386383, |
|
"eval_runtime": 2.1374, |
|
"eval_samples_per_second": 1074.669, |
|
"eval_steps_per_second": 16.843, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 0.0003506663360288954, |
|
"loss": 0.3889, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 0.00034873777607499334, |
|
"loss": 0.3888, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"eval_loss": 0.3593074679374695, |
|
"eval_runtime": 2.1457, |
|
"eval_samples_per_second": 1070.506, |
|
"eval_steps_per_second": 16.778, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 0.0003468073028890751, |
|
"loss": 0.3886, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 0.0003448750009162916, |
|
"loss": 0.3887, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"eval_loss": 0.3600543737411499, |
|
"eval_runtime": 2.1567, |
|
"eval_samples_per_second": 1065.071, |
|
"eval_steps_per_second": 16.692, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 0.00034294095468179094, |
|
"loss": 0.3881, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 0.00034100524878702073, |
|
"loss": 0.3883, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"eval_loss": 0.3610630929470062, |
|
"eval_runtime": 2.1629, |
|
"eval_samples_per_second": 1061.977, |
|
"eval_steps_per_second": 16.644, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 0.000339067967906028, |
|
"loss": 0.4695, |
|
"step": 122500 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 0.0003371291967817539, |
|
"loss": 0.6776, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"eval_loss": 0.6768860220909119, |
|
"eval_runtime": 2.1959, |
|
"eval_samples_per_second": 1046.056, |
|
"eval_steps_per_second": 16.394, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 0.0003351890202223285, |
|
"loss": 0.6171, |
|
"step": 123500 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 0.0003332475230973597, |
|
"loss": 0.3917, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"eval_loss": 0.36259856820106506, |
|
"eval_runtime": 2.1732, |
|
"eval_samples_per_second": 1056.948, |
|
"eval_steps_per_second": 16.565, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 0.00033130479033422134, |
|
"loss": 0.3889, |
|
"step": 124500 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"learning_rate": 0.0003293609069143381, |
|
"loss": 0.3897, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"eval_loss": 0.3617185652256012, |
|
"eval_runtime": 2.1669, |
|
"eval_samples_per_second": 1060.024, |
|
"eval_steps_per_second": 16.613, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 0.00032741595786946783, |
|
"loss": 0.3879, |
|
"step": 125500 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 0.000325470028277983, |
|
"loss": 0.3869, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"eval_loss": 0.35780513286590576, |
|
"eval_runtime": 2.1966, |
|
"eval_samples_per_second": 1045.71, |
|
"eval_steps_per_second": 16.389, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 0.00032352320326114754, |
|
"loss": 0.3868, |
|
"step": 126500 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 0.00032157556797939436, |
|
"loss": 0.3864, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"eval_loss": 0.35783183574676514, |
|
"eval_runtime": 2.2195, |
|
"eval_samples_per_second": 1034.94, |
|
"eval_steps_per_second": 16.22, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 0.00031962720762860057, |
|
"loss": 0.3865, |
|
"step": 127500 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 0.0003176782074363595, |
|
"loss": 0.3862, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"eval_loss": 0.35733312368392944, |
|
"eval_runtime": 2.1604, |
|
"eval_samples_per_second": 1063.224, |
|
"eval_steps_per_second": 16.664, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 0.0003157286526582535, |
|
"loss": 0.3858, |
|
"step": 128500 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 0.0003137786285741241, |
|
"loss": 0.3855, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"eval_loss": 0.3577572703361511, |
|
"eval_runtime": 2.1724, |
|
"eval_samples_per_second": 1057.343, |
|
"eval_steps_per_second": 16.571, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 0.0003118282204843421, |
|
"loss": 0.3854, |
|
"step": 129500 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 0.0003098775137060758, |
|
"loss": 0.3854, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"eval_loss": 0.357120543718338, |
|
"eval_runtime": 2.158, |
|
"eval_samples_per_second": 1064.4, |
|
"eval_steps_per_second": 16.682, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 0.00030792659356955893, |
|
"loss": 0.3853, |
|
"step": 130500 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 0.0003059755454143586, |
|
"loss": 0.3849, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"eval_loss": 0.3566107451915741, |
|
"eval_runtime": 2.2772, |
|
"eval_samples_per_second": 1008.7, |
|
"eval_steps_per_second": 15.809, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 5.87, |
|
"learning_rate": 0.00030402445458564144, |
|
"loss": 0.3845, |
|
"step": 131500 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 0.0003020734064304411, |
|
"loss": 0.3845, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"eval_loss": 0.35691574215888977, |
|
"eval_runtime": 2.2557, |
|
"eval_samples_per_second": 1018.294, |
|
"eval_steps_per_second": 15.959, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"learning_rate": 0.00030012248629392423, |
|
"loss": 0.3842, |
|
"step": 132500 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 0.00029817177951565793, |
|
"loss": 0.384, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"eval_loss": 0.35665038228034973, |
|
"eval_runtime": 2.2062, |
|
"eval_samples_per_second": 1041.164, |
|
"eval_steps_per_second": 16.318, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 0.00029622137142587594, |
|
"loss": 0.3837, |
|
"step": 133500 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 0.0002942713473417466, |
|
"loss": 0.3921, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"eval_loss": 0.3628464341163635, |
|
"eval_runtime": 2.2578, |
|
"eval_samples_per_second": 1017.356, |
|
"eval_steps_per_second": 15.945, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 0.00029232179256364054, |
|
"loss": 0.3854, |
|
"step": 134500 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 0.0002903727923713994, |
|
"loss": 0.3844, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"eval_loss": 0.3565491735935211, |
|
"eval_runtime": 2.2392, |
|
"eval_samples_per_second": 1025.813, |
|
"eval_steps_per_second": 16.077, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 0.00028842443202060556, |
|
"loss": 0.383, |
|
"step": 135500 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 0.00028647679673885255, |
|
"loss": 0.383, |
|
"step": 136000 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"eval_loss": 0.35466697812080383, |
|
"eval_runtime": 2.2168, |
|
"eval_samples_per_second": 1036.2, |
|
"eval_steps_per_second": 16.24, |
|
"step": 136000 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 0.000284529971722017, |
|
"loss": 0.3829, |
|
"step": 136500 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 0.0002825840421305321, |
|
"loss": 0.3828, |
|
"step": 137000 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"eval_loss": 0.3585571050643921, |
|
"eval_runtime": 2.2164, |
|
"eval_samples_per_second": 1036.381, |
|
"eval_steps_per_second": 16.243, |
|
"step": 137000 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 0.00028063909308566196, |
|
"loss": 0.3826, |
|
"step": 137500 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"learning_rate": 0.00027869520966577874, |
|
"loss": 0.3824, |
|
"step": 138000 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"eval_loss": 0.35530510544776917, |
|
"eval_runtime": 2.2219, |
|
"eval_samples_per_second": 1033.801, |
|
"eval_steps_per_second": 16.202, |
|
"step": 138000 |
|
}, |
|
{ |
|
"epoch": 6.18, |
|
"learning_rate": 0.00027675247690264027, |
|
"loss": 0.3823, |
|
"step": 138500 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 0.0002748109797776715, |
|
"loss": 0.3825, |
|
"step": 139000 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"eval_loss": 0.3549206256866455, |
|
"eval_runtime": 2.1769, |
|
"eval_samples_per_second": 1055.166, |
|
"eval_steps_per_second": 16.537, |
|
"step": 139000 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 0.0002728708032182461, |
|
"loss": 0.382, |
|
"step": 139500 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 0.0002709320320939721, |
|
"loss": 0.3818, |
|
"step": 140000 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"eval_loss": 0.3537048399448395, |
|
"eval_runtime": 2.2266, |
|
"eval_samples_per_second": 1031.628, |
|
"eval_steps_per_second": 16.168, |
|
"step": 140000 |
|
}, |
|
{ |
|
"epoch": 6.27, |
|
"learning_rate": 0.00026899475121297924, |
|
"loss": 0.3815, |
|
"step": 140500 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 0.00026705904531820914, |
|
"loss": 0.3815, |
|
"step": 141000 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"eval_loss": 0.3550131618976593, |
|
"eval_runtime": 2.1731, |
|
"eval_samples_per_second": 1057.017, |
|
"eval_steps_per_second": 16.566, |
|
"step": 141000 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 0.0002651249990837085, |
|
"loss": 0.3813, |
|
"step": 141500 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 0.00026319269711092485, |
|
"loss": 0.3812, |
|
"step": 142000 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"eval_loss": 0.3538789749145508, |
|
"eval_runtime": 2.1483, |
|
"eval_samples_per_second": 1069.198, |
|
"eval_steps_per_second": 16.757, |
|
"step": 142000 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 0.0002612622239250066, |
|
"loss": 0.3809, |
|
"step": 142500 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 0.0002593336639711046, |
|
"loss": 0.3806, |
|
"step": 143000 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"eval_loss": 0.35349225997924805, |
|
"eval_runtime": 2.1799, |
|
"eval_samples_per_second": 1053.718, |
|
"eval_steps_per_second": 16.515, |
|
"step": 143000 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 0.000257407101610679, |
|
"loss": 0.3805, |
|
"step": 143500 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"learning_rate": 0.00025548262111780846, |
|
"loss": 0.3804, |
|
"step": 144000 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"eval_loss": 0.3532736003398895, |
|
"eval_runtime": 2.2355, |
|
"eval_samples_per_second": 1027.515, |
|
"eval_steps_per_second": 16.104, |
|
"step": 144000 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 0.0002535603066755043, |
|
"loss": 0.3803, |
|
"step": 144500 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 0.00025164024237202764, |
|
"loss": 0.3799, |
|
"step": 145000 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"eval_loss": 0.3539290130138397, |
|
"eval_runtime": 2.1927, |
|
"eval_samples_per_second": 1047.568, |
|
"eval_steps_per_second": 16.418, |
|
"step": 145000 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 0.00024972251219721115, |
|
"loss": 0.38, |
|
"step": 145500 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 0.00024780720003878557, |
|
"loss": 0.3799, |
|
"step": 146000 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"eval_loss": 0.3527817726135254, |
|
"eval_runtime": 2.1772, |
|
"eval_samples_per_second": 1055.009, |
|
"eval_steps_per_second": 16.535, |
|
"step": 146000 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 0.00024589438967870925, |
|
"loss": 0.3795, |
|
"step": 146500 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 0.00024398416478950394, |
|
"loss": 0.3794, |
|
"step": 147000 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"eval_loss": 0.3519304692745209, |
|
"eval_runtime": 2.1956, |
|
"eval_samples_per_second": 1046.2, |
|
"eval_steps_per_second": 16.397, |
|
"step": 147000 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"learning_rate": 0.00024207660893059467, |
|
"loss": 0.3793, |
|
"step": 147500 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 0.0002401718055446543, |
|
"loss": 0.3792, |
|
"step": 148000 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"eval_loss": 0.3501195013523102, |
|
"eval_runtime": 2.1687, |
|
"eval_samples_per_second": 1059.172, |
|
"eval_steps_per_second": 16.6, |
|
"step": 148000 |
|
}, |
|
{ |
|
"epoch": 6.62, |
|
"learning_rate": 0.00023826983795395364, |
|
"loss": 0.379, |
|
"step": 148500 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 0.00023637078935671656, |
|
"loss": 0.3791, |
|
"step": 149000 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"eval_loss": 0.35130566358566284, |
|
"eval_runtime": 2.1713, |
|
"eval_samples_per_second": 1057.877, |
|
"eval_steps_per_second": 16.58, |
|
"step": 149000 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 0.00023447474282348085, |
|
"loss": 0.3787, |
|
"step": 149500 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 0.00023258178129346424, |
|
"loss": 0.3784, |
|
"step": 150000 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"eval_loss": 0.35110124945640564, |
|
"eval_runtime": 2.1647, |
|
"eval_samples_per_second": 1061.094, |
|
"eval_steps_per_second": 16.63, |
|
"step": 150000 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 0.00023069198757093631, |
|
"loss": 0.379, |
|
"step": 150500 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 0.00022880544432159663, |
|
"loss": 0.3833, |
|
"step": 151000 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"eval_loss": 0.3517898917198181, |
|
"eval_runtime": 2.2, |
|
"eval_samples_per_second": 1044.076, |
|
"eval_steps_per_second": 16.363, |
|
"step": 151000 |
|
}, |
|
{ |
|
"epoch": 6.76, |
|
"learning_rate": 0.00022692223406895848, |
|
"loss": 0.3782, |
|
"step": 151500 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 0.000225042439190739, |
|
"loss": 0.3805, |
|
"step": 152000 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"eval_loss": 0.35134270787239075, |
|
"eval_runtime": 2.2005, |
|
"eval_samples_per_second": 1043.87, |
|
"eval_steps_per_second": 16.36, |
|
"step": 152000 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 0.00022316614191525587, |
|
"loss": 0.3788, |
|
"step": 152500 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"learning_rate": 0.00022129342431783026, |
|
"loss": 0.3785, |
|
"step": 153000 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"eval_loss": 0.35215121507644653, |
|
"eval_runtime": 2.2421, |
|
"eval_samples_per_second": 1024.496, |
|
"eval_steps_per_second": 16.057, |
|
"step": 153000 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"learning_rate": 0.00021942436831719677, |
|
"loss": 0.3776, |
|
"step": 153500 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 0.00021755905567191967, |
|
"loss": 0.3772, |
|
"step": 154000 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"eval_loss": 0.34930697083473206, |
|
"eval_runtime": 2.254, |
|
"eval_samples_per_second": 1019.092, |
|
"eval_steps_per_second": 15.972, |
|
"step": 154000 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 0.00021569756797681686, |
|
"loss": 0.3775, |
|
"step": 154500 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 0.00021383998665939054, |
|
"loss": 0.3772, |
|
"step": 155000 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"eval_loss": 0.35032403469085693, |
|
"eval_runtime": 2.2005, |
|
"eval_samples_per_second": 1043.877, |
|
"eval_steps_per_second": 16.36, |
|
"step": 155000 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 0.00021198639297626516, |
|
"loss": 0.3769, |
|
"step": 155500 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 0.0002101368680096334, |
|
"loss": 0.3771, |
|
"step": 156000 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"eval_loss": 0.35130995512008667, |
|
"eval_runtime": 2.2636, |
|
"eval_samples_per_second": 1014.77, |
|
"eval_steps_per_second": 15.904, |
|
"step": 156000 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 0.00020829149266370862, |
|
"loss": 0.377, |
|
"step": 156500 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 0.00020645034766118703, |
|
"loss": 0.3769, |
|
"step": 157000 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.35048824548721313, |
|
"eval_runtime": 2.2749, |
|
"eval_samples_per_second": 1009.702, |
|
"eval_steps_per_second": 15.825, |
|
"step": 157000 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 0.00020461351353971526, |
|
"loss": 0.3766, |
|
"step": 157500 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 0.00020278107064836847, |
|
"loss": 0.3766, |
|
"step": 158000 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"eval_loss": 0.3498741090297699, |
|
"eval_runtime": 2.2625, |
|
"eval_samples_per_second": 1015.239, |
|
"eval_steps_per_second": 15.911, |
|
"step": 158000 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 0.00020095309914413485, |
|
"loss": 0.3764, |
|
"step": 158500 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 0.00019912967898840997, |
|
"loss": 0.3762, |
|
"step": 159000 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"eval_loss": 0.34895479679107666, |
|
"eval_runtime": 2.1885, |
|
"eval_samples_per_second": 1049.579, |
|
"eval_steps_per_second": 16.45, |
|
"step": 159000 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 0.00019731088994349834, |
|
"loss": 0.3763, |
|
"step": 159500 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"learning_rate": 0.0001954968115691248, |
|
"loss": 0.376, |
|
"step": 160000 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"eval_loss": 0.3464561402797699, |
|
"eval_runtime": 2.2955, |
|
"eval_samples_per_second": 1000.637, |
|
"eval_steps_per_second": 15.683, |
|
"step": 160000 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 0.00019368752321895452, |
|
"loss": 0.3758, |
|
"step": 160500 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 0.00019188310403712105, |
|
"loss": 0.3756, |
|
"step": 161000 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"eval_loss": 0.3490326702594757, |
|
"eval_runtime": 2.2155, |
|
"eval_samples_per_second": 1036.763, |
|
"eval_steps_per_second": 16.249, |
|
"step": 161000 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 0.00019008363295476495, |
|
"loss": 0.3756, |
|
"step": 161500 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 0.0001882891886865808, |
|
"loss": 0.3753, |
|
"step": 162000 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"eval_loss": 0.3482523560523987, |
|
"eval_runtime": 2.2246, |
|
"eval_samples_per_second": 1032.556, |
|
"eval_steps_per_second": 16.183, |
|
"step": 162000 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 0.00018649984972737404, |
|
"loss": 0.3757, |
|
"step": 162500 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 0.00018471569434862749, |
|
"loss": 0.3749, |
|
"step": 163000 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"eval_loss": 0.3480832874774933, |
|
"eval_runtime": 2.1706, |
|
"eval_samples_per_second": 1058.217, |
|
"eval_steps_per_second": 16.585, |
|
"step": 163000 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 0.00018293680059507713, |
|
"loss": 0.3753, |
|
"step": 163500 |
|
}, |
|
{ |
|
"epoch": 7.32, |
|
"learning_rate": 0.00018116324628129882, |
|
"loss": 0.3747, |
|
"step": 164000 |
|
}, |
|
{ |
|
"epoch": 7.32, |
|
"eval_loss": 0.3469808101654053, |
|
"eval_runtime": 2.1769, |
|
"eval_samples_per_second": 1055.189, |
|
"eval_steps_per_second": 16.538, |
|
"step": 164000 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 0.00017939510898830357, |
|
"loss": 0.3746, |
|
"step": 164500 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 0.0001776324660601446, |
|
"loss": 0.375, |
|
"step": 165000 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"eval_loss": 0.3475760519504547, |
|
"eval_runtime": 2.1442, |
|
"eval_samples_per_second": 1071.239, |
|
"eval_steps_per_second": 16.789, |
|
"step": 165000 |
|
}, |
|
{ |
|
"epoch": 7.38, |
|
"learning_rate": 0.00017587539460053368, |
|
"loss": 0.3743, |
|
"step": 165500 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 0.0001741239714694688, |
|
"loss": 0.3742, |
|
"step": 166000 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"eval_loss": 0.34710729122161865, |
|
"eval_runtime": 2.2456, |
|
"eval_samples_per_second": 1022.894, |
|
"eval_steps_per_second": 16.031, |
|
"step": 166000 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 0.0001723782732798713, |
|
"loss": 0.3741, |
|
"step": 166500 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 0.00017063837639423517, |
|
"loss": 0.3741, |
|
"step": 167000 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"eval_loss": 0.3461616337299347, |
|
"eval_runtime": 2.2634, |
|
"eval_samples_per_second": 1014.851, |
|
"eval_steps_per_second": 15.905, |
|
"step": 167000 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 0.00016890435692128712, |
|
"loss": 0.3738, |
|
"step": 167500 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 0.0001671762907126559, |
|
"loss": 0.3738, |
|
"step": 168000 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"eval_loss": 0.3469841480255127, |
|
"eval_runtime": 2.2202, |
|
"eval_samples_per_second": 1034.588, |
|
"eval_steps_per_second": 16.215, |
|
"step": 168000 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 0.00016545425335955596, |
|
"loss": 0.3735, |
|
"step": 168500 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"learning_rate": 0.00016373832018947945, |
|
"loss": 0.3735, |
|
"step": 169000 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"eval_loss": 0.3462463617324829, |
|
"eval_runtime": 2.2185, |
|
"eval_samples_per_second": 1035.373, |
|
"eval_steps_per_second": 16.227, |
|
"step": 169000 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"learning_rate": 0.0001620285662629024, |
|
"loss": 0.3735, |
|
"step": 169500 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 0.0001603250663700002, |
|
"loss": 0.3736, |
|
"step": 170000 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"eval_loss": 0.3466815948486328, |
|
"eval_runtime": 2.223, |
|
"eval_samples_per_second": 1033.305, |
|
"eval_steps_per_second": 16.195, |
|
"step": 170000 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 0.00015862789502737648, |
|
"loss": 0.3728, |
|
"step": 170500 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"learning_rate": 0.00015693712647480446, |
|
"loss": 0.3731, |
|
"step": 171000 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"eval_loss": 0.3457169234752655, |
|
"eval_runtime": 2.2902, |
|
"eval_samples_per_second": 1002.962, |
|
"eval_steps_per_second": 15.719, |
|
"step": 171000 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"learning_rate": 0.00015525283467197743, |
|
"loss": 0.3727, |
|
"step": 171500 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"learning_rate": 0.00015357509329527556, |
|
"loss": 0.3726, |
|
"step": 172000 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"eval_loss": 0.3477800190448761, |
|
"eval_runtime": 2.185, |
|
"eval_samples_per_second": 1051.256, |
|
"eval_steps_per_second": 16.476, |
|
"step": 172000 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 0.00015190397573454158, |
|
"loss": 0.3727, |
|
"step": 172500 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 0.00015023955508987127, |
|
"loss": 0.3725, |
|
"step": 173000 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"eval_loss": 0.3447197675704956, |
|
"eval_runtime": 2.1651, |
|
"eval_samples_per_second": 1060.935, |
|
"eval_steps_per_second": 16.628, |
|
"step": 173000 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 0.00014858190416841565, |
|
"loss": 0.3724, |
|
"step": 173500 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 0.00014693109548119591, |
|
"loss": 0.3722, |
|
"step": 174000 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"eval_loss": 0.3459009826183319, |
|
"eval_runtime": 2.2671, |
|
"eval_samples_per_second": 1013.172, |
|
"eval_steps_per_second": 15.879, |
|
"step": 174000 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 0.00014528720123993226, |
|
"loss": 0.3721, |
|
"step": 174500 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"learning_rate": 0.0001436502933538841, |
|
"loss": 0.3723, |
|
"step": 175000 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"eval_loss": 0.3462165296077728, |
|
"eval_runtime": 2.2195, |
|
"eval_samples_per_second": 1034.906, |
|
"eval_steps_per_second": 16.22, |
|
"step": 175000 |
|
}, |
|
{ |
|
"epoch": 7.83, |
|
"learning_rate": 0.00014202044342670508, |
|
"loss": 0.372, |
|
"step": 175500 |
|
}, |
|
{ |
|
"epoch": 7.85, |
|
"learning_rate": 0.00014039772275331125, |
|
"loss": 0.3718, |
|
"step": 176000 |
|
}, |
|
{ |
|
"epoch": 7.85, |
|
"eval_loss": 0.3463585674762726, |
|
"eval_runtime": 2.2454, |
|
"eval_samples_per_second": 1022.987, |
|
"eval_steps_per_second": 16.033, |
|
"step": 176000 |
|
}, |
|
{ |
|
"epoch": 7.87, |
|
"learning_rate": 0.00013878220231676152, |
|
"loss": 0.3716, |
|
"step": 176500 |
|
}, |
|
{ |
|
"epoch": 7.9, |
|
"learning_rate": 0.00013717395278515355, |
|
"loss": 0.3716, |
|
"step": 177000 |
|
}, |
|
{ |
|
"epoch": 7.9, |
|
"eval_loss": 0.34527209401130676, |
|
"eval_runtime": 2.2682, |
|
"eval_samples_per_second": 1012.694, |
|
"eval_steps_per_second": 15.872, |
|
"step": 177000 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"learning_rate": 0.00013557304450853162, |
|
"loss": 0.3714, |
|
"step": 177500 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 0.00013397954751581014, |
|
"loss": 0.3712, |
|
"step": 178000 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"eval_loss": 0.34656643867492676, |
|
"eval_runtime": 2.2489, |
|
"eval_samples_per_second": 1021.386, |
|
"eval_steps_per_second": 16.008, |
|
"step": 178000 |
|
}, |
|
{ |
|
"epoch": 7.96, |
|
"learning_rate": 0.00013239353151170983, |
|
"loss": 0.371, |
|
"step": 178500 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 0.00013081506587370853, |
|
"loss": 0.3712, |
|
"step": 179000 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"eval_loss": 0.34555310010910034, |
|
"eval_runtime": 2.1617, |
|
"eval_samples_per_second": 1062.607, |
|
"eval_steps_per_second": 16.654, |
|
"step": 179000 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 0.00012924421964900695, |
|
"loss": 0.371, |
|
"step": 179500 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 0.00012768106155150758, |
|
"loss": 0.3709, |
|
"step": 180000 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"eval_loss": 0.34523507952690125, |
|
"eval_runtime": 2.3269, |
|
"eval_samples_per_second": 987.138, |
|
"eval_steps_per_second": 15.471, |
|
"step": 180000 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"learning_rate": 0.00012612565995880976, |
|
"loss": 0.3725, |
|
"step": 180500 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"learning_rate": 0.00012457808290921774, |
|
"loss": 0.3709, |
|
"step": 181000 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"eval_loss": 0.3426642119884491, |
|
"eval_runtime": 2.2304, |
|
"eval_samples_per_second": 1029.848, |
|
"eval_steps_per_second": 16.14, |
|
"step": 181000 |
|
}, |
|
{ |
|
"epoch": 8.1, |
|
"learning_rate": 0.00012303839809876525, |
|
"loss": 0.3705, |
|
"step": 181500 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 0.00012150667287825382, |
|
"loss": 0.3707, |
|
"step": 182000 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"eval_loss": 0.3444533050060272, |
|
"eval_runtime": 2.1579, |
|
"eval_samples_per_second": 1064.458, |
|
"eval_steps_per_second": 16.683, |
|
"step": 182000 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"learning_rate": 0.00011998297425030656, |
|
"loss": 0.3705, |
|
"step": 182500 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 0.00011846736886643775, |
|
"loss": 0.3703, |
|
"step": 183000 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"eval_loss": 0.34523776173591614, |
|
"eval_runtime": 2.2575, |
|
"eval_samples_per_second": 1017.476, |
|
"eval_steps_per_second": 15.947, |
|
"step": 183000 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 0.00011695992302413651, |
|
"loss": 0.3702, |
|
"step": 183500 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"learning_rate": 0.00011546070266396771, |
|
"loss": 0.3701, |
|
"step": 184000 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"eval_loss": 0.34200483560562134, |
|
"eval_runtime": 2.2775, |
|
"eval_samples_per_second": 1008.549, |
|
"eval_steps_per_second": 15.807, |
|
"step": 184000 |
|
}, |
|
{ |
|
"epoch": 8.23, |
|
"learning_rate": 0.00011396977336668645, |
|
"loss": 0.3701, |
|
"step": 184500 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"learning_rate": 0.00011248720035037021, |
|
"loss": 0.3699, |
|
"step": 185000 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"eval_loss": 0.34288346767425537, |
|
"eval_runtime": 2.2113, |
|
"eval_samples_per_second": 1038.766, |
|
"eval_steps_per_second": 16.28, |
|
"step": 185000 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"learning_rate": 0.00011101304846756577, |
|
"loss": 0.3699, |
|
"step": 185500 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 0.00010954738220245183, |
|
"loss": 0.3697, |
|
"step": 186000 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"eval_loss": 0.3431689739227295, |
|
"eval_runtime": 2.222, |
|
"eval_samples_per_second": 1033.767, |
|
"eval_steps_per_second": 16.202, |
|
"step": 186000 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"learning_rate": 0.00010809026566801912, |
|
"loss": 0.3695, |
|
"step": 186500 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 0.00010664176260326507, |
|
"loss": 0.3696, |
|
"step": 187000 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"eval_loss": 0.3424961268901825, |
|
"eval_runtime": 2.2738, |
|
"eval_samples_per_second": 1010.209, |
|
"eval_steps_per_second": 15.833, |
|
"step": 187000 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 0.00010520193637040641, |
|
"loss": 0.3697, |
|
"step": 187500 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"learning_rate": 0.00010377084995210682, |
|
"loss": 0.3696, |
|
"step": 188000 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"eval_loss": 0.34366557002067566, |
|
"eval_runtime": 2.2531, |
|
"eval_samples_per_second": 1019.467, |
|
"eval_steps_per_second": 15.978, |
|
"step": 188000 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 0.00010234856594872234, |
|
"loss": 0.3692, |
|
"step": 188500 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 0.00010093514657556295, |
|
"loss": 0.3694, |
|
"step": 189000 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"eval_loss": 0.34253063797950745, |
|
"eval_runtime": 2.2641, |
|
"eval_samples_per_second": 1014.519, |
|
"eval_steps_per_second": 15.9, |
|
"step": 189000 |
|
}, |
|
{ |
|
"epoch": 8.45, |
|
"learning_rate": 9.953065366017073e-05, |
|
"loss": 0.369, |
|
"step": 189500 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"learning_rate": 9.813514863961586e-05, |
|
"loss": 0.369, |
|
"step": 190000 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"eval_loss": 0.34287169575691223, |
|
"eval_runtime": 2.2268, |
|
"eval_samples_per_second": 1031.502, |
|
"eval_steps_per_second": 16.166, |
|
"step": 190000 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 9.67486925578087e-05, |
|
"loss": 0.3686, |
|
"step": 190500 |
|
}, |
|
{ |
|
"epoch": 8.52, |
|
"learning_rate": 9.537134606282964e-05, |
|
"loss": 0.369, |
|
"step": 191000 |
|
}, |
|
{ |
|
"epoch": 8.52, |
|
"eval_loss": 0.3414945900440216, |
|
"eval_runtime": 2.186, |
|
"eval_samples_per_second": 1050.801, |
|
"eval_steps_per_second": 16.469, |
|
"step": 191000 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"learning_rate": 9.400316940427652e-05, |
|
"loss": 0.3688, |
|
"step": 191500 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 9.264422243062844e-05, |
|
"loss": 0.3685, |
|
"step": 192000 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"eval_loss": 0.34313786029815674, |
|
"eval_runtime": 2.2429, |
|
"eval_samples_per_second": 1024.133, |
|
"eval_steps_per_second": 16.051, |
|
"step": 192000 |
|
}, |
|
{ |
|
"epoch": 8.59, |
|
"learning_rate": 9.129456458662876e-05, |
|
"loss": 0.3684, |
|
"step": 192500 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"learning_rate": 8.995425491068365e-05, |
|
"loss": 0.3684, |
|
"step": 193000 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"eval_loss": 0.341512531042099, |
|
"eval_runtime": 2.1692, |
|
"eval_samples_per_second": 1058.928, |
|
"eval_steps_per_second": 16.596, |
|
"step": 193000 |
|
}, |
|
{ |
|
"epoch": 8.63, |
|
"learning_rate": 8.862335203228025e-05, |
|
"loss": 0.3683, |
|
"step": 193500 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"learning_rate": 8.73019141694222e-05, |
|
"loss": 0.3683, |
|
"step": 194000 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"eval_loss": 0.3420671820640564, |
|
"eval_runtime": 2.2415, |
|
"eval_samples_per_second": 1024.765, |
|
"eval_steps_per_second": 16.061, |
|
"step": 194000 |
|
}, |
|
{ |
|
"epoch": 8.68, |
|
"learning_rate": 8.598999912608229e-05, |
|
"loss": 0.3681, |
|
"step": 194500 |
|
}, |
|
{ |
|
"epoch": 8.7, |
|
"learning_rate": 8.468766428967468e-05, |
|
"loss": 0.368, |
|
"step": 195000 |
|
}, |
|
{ |
|
"epoch": 8.7, |
|
"eval_loss": 0.3422170877456665, |
|
"eval_runtime": 2.2216, |
|
"eval_samples_per_second": 1033.959, |
|
"eval_steps_per_second": 16.205, |
|
"step": 195000 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 8.339496662854397e-05, |
|
"loss": 0.3682, |
|
"step": 195500 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"learning_rate": 8.211196268947367e-05, |
|
"loss": 0.3719, |
|
"step": 196000 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"eval_loss": 0.3433374762535095, |
|
"eval_runtime": 2.2305, |
|
"eval_samples_per_second": 1029.81, |
|
"eval_steps_per_second": 16.14, |
|
"step": 196000 |
|
}, |
|
{ |
|
"epoch": 8.77, |
|
"learning_rate": 8.083870859521251e-05, |
|
"loss": 0.368, |
|
"step": 196500 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 7.95752600420192e-05, |
|
"loss": 0.3678, |
|
"step": 197000 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"eval_loss": 0.33997857570648193, |
|
"eval_runtime": 2.2361, |
|
"eval_samples_per_second": 1027.215, |
|
"eval_steps_per_second": 16.099, |
|
"step": 197000 |
|
}, |
|
{ |
|
"epoch": 8.81, |
|
"learning_rate": 7.832167229722666e-05, |
|
"loss": 0.3677, |
|
"step": 197500 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 7.707800019682362e-05, |
|
"loss": 0.3675, |
|
"step": 198000 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"eval_loss": 0.3419722616672516, |
|
"eval_runtime": 2.2466, |
|
"eval_samples_per_second": 1022.412, |
|
"eval_steps_per_second": 16.024, |
|
"step": 198000 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 7.5844298143057e-05, |
|
"loss": 0.3676, |
|
"step": 198500 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 7.462062010205106e-05, |
|
"loss": 0.3676, |
|
"step": 199000 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"eval_loss": 0.3425526022911072, |
|
"eval_runtime": 2.2092, |
|
"eval_samples_per_second": 1039.756, |
|
"eval_steps_per_second": 16.296, |
|
"step": 199000 |
|
}, |
|
{ |
|
"epoch": 8.9, |
|
"learning_rate": 7.340701960144751e-05, |
|
"loss": 0.3676, |
|
"step": 199500 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"learning_rate": 7.220354972806392e-05, |
|
"loss": 0.3674, |
|
"step": 200000 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"eval_loss": 0.33957439661026, |
|
"eval_runtime": 2.28, |
|
"eval_samples_per_second": 1007.437, |
|
"eval_steps_per_second": 15.789, |
|
"step": 200000 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 7.101026312557103e-05, |
|
"loss": 0.3672, |
|
"step": 200500 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 6.982721199219075e-05, |
|
"loss": 0.3673, |
|
"step": 201000 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"eval_loss": 0.3403836786746979, |
|
"eval_runtime": 2.153, |
|
"eval_samples_per_second": 1066.9, |
|
"eval_steps_per_second": 16.721, |
|
"step": 201000 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 6.865444807841203e-05, |
|
"loss": 0.3671, |
|
"step": 201500 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.749202268472787e-05, |
|
"loss": 0.3671, |
|
"step": 202000 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"eval_loss": 0.3397007882595062, |
|
"eval_runtime": 2.2258, |
|
"eval_samples_per_second": 1032.002, |
|
"eval_steps_per_second": 16.174, |
|
"step": 202000 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 6.633998665939053e-05, |
|
"loss": 0.3672, |
|
"step": 202500 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 6.519839039618793e-05, |
|
"loss": 0.3669, |
|
"step": 203000 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"eval_loss": 0.3417140245437622, |
|
"eval_runtime": 2.2137, |
|
"eval_samples_per_second": 1037.631, |
|
"eval_steps_per_second": 16.262, |
|
"step": 203000 |
|
}, |
|
{ |
|
"epoch": 9.08, |
|
"learning_rate": 6.406728383223897e-05, |
|
"loss": 0.3669, |
|
"step": 203500 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 6.294671644580888e-05, |
|
"loss": 0.3669, |
|
"step": 204000 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"eval_loss": 0.3413463830947876, |
|
"eval_runtime": 2.179, |
|
"eval_samples_per_second": 1054.147, |
|
"eval_steps_per_second": 16.521, |
|
"step": 204000 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 6.18367372541454e-05, |
|
"loss": 0.3668, |
|
"step": 204500 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"learning_rate": 6.073739481133408e-05, |
|
"loss": 0.3666, |
|
"step": 205000 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"eval_loss": 0.33860939741134644, |
|
"eval_runtime": 2.2027, |
|
"eval_samples_per_second": 1042.794, |
|
"eval_steps_per_second": 16.343, |
|
"step": 205000 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 5.96487372061749e-05, |
|
"loss": 0.3667, |
|
"step": 205500 |
|
}, |
|
{ |
|
"epoch": 9.19, |
|
"learning_rate": 5.857081206007811e-05, |
|
"loss": 0.3666, |
|
"step": 206000 |
|
}, |
|
{ |
|
"epoch": 9.19, |
|
"eval_loss": 0.3413560390472412, |
|
"eval_runtime": 2.2133, |
|
"eval_samples_per_second": 1037.836, |
|
"eval_steps_per_second": 16.266, |
|
"step": 206000 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 5.750366652498169e-05, |
|
"loss": 0.3663, |
|
"step": 206500 |
|
}, |
|
{ |
|
"epoch": 9.23, |
|
"learning_rate": 5.6447347281288474e-05, |
|
"loss": 0.3664, |
|
"step": 207000 |
|
}, |
|
{ |
|
"epoch": 9.23, |
|
"eval_loss": 0.3407277464866638, |
|
"eval_runtime": 2.2449, |
|
"eval_samples_per_second": 1023.218, |
|
"eval_steps_per_second": 16.037, |
|
"step": 207000 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 5.540190053582401e-05, |
|
"loss": 0.3665, |
|
"step": 207500 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"learning_rate": 5.43673720198159e-05, |
|
"loss": 0.3662, |
|
"step": 208000 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"eval_loss": 0.3401471674442291, |
|
"eval_runtime": 2.2056, |
|
"eval_samples_per_second": 1041.45, |
|
"eval_steps_per_second": 16.322, |
|
"step": 208000 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 5.33438069868928e-05, |
|
"loss": 0.3662, |
|
"step": 208500 |
|
}, |
|
{ |
|
"epoch": 9.32, |
|
"learning_rate": 5.2331250211105e-05, |
|
"loss": 0.3661, |
|
"step": 209000 |
|
}, |
|
{ |
|
"epoch": 9.32, |
|
"eval_loss": 0.34120145440101624, |
|
"eval_runtime": 2.1294, |
|
"eval_samples_per_second": 1078.699, |
|
"eval_steps_per_second": 16.906, |
|
"step": 209000 |
|
}, |
|
{ |
|
"epoch": 9.35, |
|
"learning_rate": 5.13297459849662e-05, |
|
"loss": 0.366, |
|
"step": 209500 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 5.0339338117515696e-05, |
|
"loss": 0.366, |
|
"step": 210000 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"eval_loss": 0.3374158442020416, |
|
"eval_runtime": 2.1871, |
|
"eval_samples_per_second": 1050.236, |
|
"eval_steps_per_second": 16.46, |
|
"step": 210000 |
|
}, |
|
{ |
|
"epoch": 9.39, |
|
"learning_rate": 4.9360069932402115e-05, |
|
"loss": 0.3659, |
|
"step": 210500 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 4.839198426598824e-05, |
|
"loss": 0.3659, |
|
"step": 211000 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"eval_loss": 0.3399938941001892, |
|
"eval_runtime": 2.2613, |
|
"eval_samples_per_second": 1015.803, |
|
"eval_steps_per_second": 15.92, |
|
"step": 211000 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 4.7435123465477156e-05, |
|
"loss": 0.3659, |
|
"step": 211500 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 4.6489529387060306e-05, |
|
"loss": 0.3658, |
|
"step": 212000 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"eval_loss": 0.34058016538619995, |
|
"eval_runtime": 2.2213, |
|
"eval_samples_per_second": 1034.088, |
|
"eval_steps_per_second": 16.207, |
|
"step": 212000 |
|
}, |
|
{ |
|
"epoch": 9.48, |
|
"learning_rate": 4.555524339408575e-05, |
|
"loss": 0.3658, |
|
"step": 212500 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 4.46323063552496e-05, |
|
"loss": 0.3658, |
|
"step": 213000 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"eval_loss": 0.33832791447639465, |
|
"eval_runtime": 2.1746, |
|
"eval_samples_per_second": 1056.275, |
|
"eval_steps_per_second": 16.555, |
|
"step": 213000 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 4.3720758642807817e-05, |
|
"loss": 0.3655, |
|
"step": 213500 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 4.2820640130810446e-05, |
|
"loss": 0.3656, |
|
"step": 214000 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"eval_loss": 0.33994585275650024, |
|
"eval_runtime": 2.2047, |
|
"eval_samples_per_second": 1041.878, |
|
"eval_steps_per_second": 16.329, |
|
"step": 214000 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 4.193199019335709e-05, |
|
"loss": 0.3654, |
|
"step": 214500 |
|
}, |
|
{ |
|
"epoch": 9.59, |
|
"learning_rate": 4.105484770287477e-05, |
|
"loss": 0.3655, |
|
"step": 215000 |
|
}, |
|
{ |
|
"epoch": 9.59, |
|
"eval_loss": 0.3385065793991089, |
|
"eval_runtime": 2.1784, |
|
"eval_samples_per_second": 1054.433, |
|
"eval_steps_per_second": 16.526, |
|
"step": 215000 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 4.018925102841773e-05, |
|
"loss": 0.3652, |
|
"step": 215500 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 3.9335238033988484e-05, |
|
"loss": 0.3653, |
|
"step": 216000 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"eval_loss": 0.3406268358230591, |
|
"eval_runtime": 2.2102, |
|
"eval_samples_per_second": 1039.279, |
|
"eval_steps_per_second": 16.288, |
|
"step": 216000 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 3.8492846076882115e-05, |
|
"loss": 0.3652, |
|
"step": 216500 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 3.766211200605186e-05, |
|
"loss": 0.3652, |
|
"step": 217000 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"eval_loss": 0.33878666162490845, |
|
"eval_runtime": 2.1865, |
|
"eval_samples_per_second": 1050.535, |
|
"eval_steps_per_second": 16.465, |
|
"step": 217000 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 3.684307216049706e-05, |
|
"loss": 0.3653, |
|
"step": 217500 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 3.6035762367673984e-05, |
|
"loss": 0.3674, |
|
"step": 218000 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"eval_loss": 0.33809101581573486, |
|
"eval_runtime": 2.1832, |
|
"eval_samples_per_second": 1052.106, |
|
"eval_steps_per_second": 16.489, |
|
"step": 218000 |
|
}, |
|
{ |
|
"epoch": 9.75, |
|
"learning_rate": 3.52402179419282e-05, |
|
"loss": 0.3649, |
|
"step": 218500 |
|
}, |
|
{ |
|
"epoch": 9.77, |
|
"learning_rate": 3.4456473682950194e-05, |
|
"loss": 0.365, |
|
"step": 219000 |
|
}, |
|
{ |
|
"epoch": 9.77, |
|
"eval_loss": 0.33866041898727417, |
|
"eval_runtime": 2.2773, |
|
"eval_samples_per_second": 1008.633, |
|
"eval_steps_per_second": 15.808, |
|
"step": 219000 |
|
}, |
|
{ |
|
"epoch": 9.79, |
|
"learning_rate": 3.3684563874252695e-05, |
|
"loss": 0.365, |
|
"step": 219500 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"learning_rate": 3.2924522281671496e-05, |
|
"loss": 0.3648, |
|
"step": 220000 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"eval_loss": 0.33735984563827515, |
|
"eval_runtime": 2.208, |
|
"eval_samples_per_second": 1040.301, |
|
"eval_steps_per_second": 16.304, |
|
"step": 220000 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 3.2176382151888054e-05, |
|
"loss": 0.365, |
|
"step": 220500 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 3.1440176210975204e-05, |
|
"loss": 0.3649, |
|
"step": 221000 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"eval_loss": 0.3377821445465088, |
|
"eval_runtime": 2.2679, |
|
"eval_samples_per_second": 1012.829, |
|
"eval_steps_per_second": 15.874, |
|
"step": 221000 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 3.071593666296585e-05, |
|
"loss": 0.3648, |
|
"step": 221500 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 3.000369518844396e-05, |
|
"loss": 0.3649, |
|
"step": 222000 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"eval_loss": 0.33791235089302063, |
|
"eval_runtime": 2.2027, |
|
"eval_samples_per_second": 1042.832, |
|
"eval_steps_per_second": 16.344, |
|
"step": 222000 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 2.9303482943159077e-05, |
|
"loss": 0.3648, |
|
"step": 222500 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"learning_rate": 2.861533055666306e-05, |
|
"loss": 0.3646, |
|
"step": 223000 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"eval_loss": 0.33819380402565, |
|
"eval_runtime": 2.1945, |
|
"eval_samples_per_second": 1046.691, |
|
"eval_steps_per_second": 16.404, |
|
"step": 223000 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 2.793926813097066e-05, |
|
"loss": 0.3645, |
|
"step": 223500 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"learning_rate": 2.7275325239242546e-05, |
|
"loss": 0.3647, |
|
"step": 224000 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"eval_loss": 0.33774814009666443, |
|
"eval_runtime": 2.2427, |
|
"eval_samples_per_second": 1024.194, |
|
"eval_steps_per_second": 16.052, |
|
"step": 224000 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 2.6623530924491626e-05, |
|
"loss": 0.3645, |
|
"step": 224500 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 2.5983913698312782e-05, |
|
"loss": 0.3644, |
|
"step": 225000 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"eval_loss": 0.33505749702453613, |
|
"eval_runtime": 2.2238, |
|
"eval_samples_per_second": 1032.916, |
|
"eval_steps_per_second": 16.188, |
|
"step": 225000 |
|
}, |
|
{ |
|
"epoch": 10.06, |
|
"learning_rate": 2.5356501539635512e-05, |
|
"loss": 0.3644, |
|
"step": 225500 |
|
}, |
|
{ |
|
"epoch": 10.08, |
|
"learning_rate": 2.4741321893500244e-05, |
|
"loss": 0.3644, |
|
"step": 226000 |
|
}, |
|
{ |
|
"epoch": 10.08, |
|
"eval_loss": 0.337401807308197, |
|
"eval_runtime": 2.2527, |
|
"eval_samples_per_second": 1019.654, |
|
"eval_steps_per_second": 15.981, |
|
"step": 226000 |
|
}, |
|
{ |
|
"epoch": 10.1, |
|
"learning_rate": 2.4138401669857587e-05, |
|
"loss": 0.3644, |
|
"step": 226500 |
|
}, |
|
{ |
|
"epoch": 10.13, |
|
"learning_rate": 2.3547767242391212e-05, |
|
"loss": 0.3644, |
|
"step": 227000 |
|
}, |
|
{ |
|
"epoch": 10.13, |
|
"eval_loss": 0.33791208267211914, |
|
"eval_runtime": 2.1965, |
|
"eval_samples_per_second": 1045.741, |
|
"eval_steps_per_second": 16.389, |
|
"step": 227000 |
|
}, |
|
{ |
|
"epoch": 10.15, |
|
"learning_rate": 2.2969444447364498e-05, |
|
"loss": 0.3642, |
|
"step": 227500 |
|
}, |
|
{ |
|
"epoch": 10.17, |
|
"learning_rate": 2.240345858248992e-05, |
|
"loss": 0.3651, |
|
"step": 228000 |
|
}, |
|
{ |
|
"epoch": 10.17, |
|
"eval_loss": 0.33648931980133057, |
|
"eval_runtime": 2.2235, |
|
"eval_samples_per_second": 1033.033, |
|
"eval_steps_per_second": 16.19, |
|
"step": 228000 |
|
}, |
|
{ |
|
"epoch": 10.19, |
|
"learning_rate": 2.184983440582284e-05, |
|
"loss": 0.3644, |
|
"step": 228500 |
|
}, |
|
{ |
|
"epoch": 10.22, |
|
"learning_rate": 2.1308596134678134e-05, |
|
"loss": 0.3643, |
|
"step": 229000 |
|
}, |
|
{ |
|
"epoch": 10.22, |
|
"eval_loss": 0.3360297977924347, |
|
"eval_runtime": 2.2193, |
|
"eval_samples_per_second": 1035.0, |
|
"eval_steps_per_second": 16.221, |
|
"step": 229000 |
|
}, |
|
{ |
|
"epoch": 10.24, |
|
"learning_rate": 2.0779767444571236e-05, |
|
"loss": 0.3643, |
|
"step": 229500 |
|
}, |
|
{ |
|
"epoch": 10.26, |
|
"learning_rate": 2.0263371468182175e-05, |
|
"loss": 0.3642, |
|
"step": 230000 |
|
}, |
|
{ |
|
"epoch": 10.26, |
|
"eval_loss": 0.33706384897232056, |
|
"eval_runtime": 2.2761, |
|
"eval_samples_per_second": 1009.19, |
|
"eval_steps_per_second": 15.817, |
|
"step": 230000 |
|
}, |
|
{ |
|
"epoch": 10.28, |
|
"learning_rate": 1.975943079434381e-05, |
|
"loss": 0.3642, |
|
"step": 230500 |
|
}, |
|
{ |
|
"epoch": 10.31, |
|
"learning_rate": 1.9267967467053834e-05, |
|
"loss": 0.364, |
|
"step": 231000 |
|
}, |
|
{ |
|
"epoch": 10.31, |
|
"eval_loss": 0.33804792165756226, |
|
"eval_runtime": 2.1938, |
|
"eval_samples_per_second": 1047.056, |
|
"eval_steps_per_second": 16.41, |
|
"step": 231000 |
|
}, |
|
{ |
|
"epoch": 10.33, |
|
"learning_rate": 1.878900298451024e-05, |
|
"loss": 0.3639, |
|
"step": 231500 |
|
}, |
|
{ |
|
"epoch": 10.35, |
|
"learning_rate": 1.83225582981712e-05, |
|
"loss": 0.364, |
|
"step": 232000 |
|
}, |
|
{ |
|
"epoch": 10.35, |
|
"eval_loss": 0.3375454843044281, |
|
"eval_runtime": 2.2078, |
|
"eval_samples_per_second": 1040.381, |
|
"eval_steps_per_second": 16.305, |
|
"step": 232000 |
|
}, |
|
{ |
|
"epoch": 10.37, |
|
"learning_rate": 1.7868653811838307e-05, |
|
"loss": 0.3638, |
|
"step": 232500 |
|
}, |
|
{ |
|
"epoch": 10.39, |
|
"learning_rate": 1.7427309380764227e-05, |
|
"loss": 0.364, |
|
"step": 233000 |
|
}, |
|
{ |
|
"epoch": 10.39, |
|
"eval_loss": 0.33860087394714355, |
|
"eval_runtime": 2.2, |
|
"eval_samples_per_second": 1044.098, |
|
"eval_steps_per_second": 16.364, |
|
"step": 233000 |
|
}, |
|
{ |
|
"epoch": 10.42, |
|
"learning_rate": 1.6998544310784175e-05, |
|
"loss": 0.3638, |
|
"step": 233500 |
|
}, |
|
{ |
|
"epoch": 10.44, |
|
"learning_rate": 1.6582377357471285e-05, |
|
"loss": 0.3639, |
|
"step": 234000 |
|
}, |
|
{ |
|
"epoch": 10.44, |
|
"eval_loss": 0.33734622597694397, |
|
"eval_runtime": 2.1793, |
|
"eval_samples_per_second": 1054.021, |
|
"eval_steps_per_second": 16.519, |
|
"step": 234000 |
|
}, |
|
{ |
|
"epoch": 10.46, |
|
"learning_rate": 1.617882672531633e-05, |
|
"loss": 0.3639, |
|
"step": 234500 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"learning_rate": 1.578791006693124e-05, |
|
"loss": 0.364, |
|
"step": 235000 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"eval_loss": 0.3377144932746887, |
|
"eval_runtime": 2.2337, |
|
"eval_samples_per_second": 1028.33, |
|
"eval_steps_per_second": 16.117, |
|
"step": 235000 |
|
}, |
|
{ |
|
"epoch": 10.51, |
|
"learning_rate": 1.5409644482277073e-05, |
|
"loss": 0.3638, |
|
"step": 235500 |
|
}, |
|
{ |
|
"epoch": 10.53, |
|
"learning_rate": 1.504404651791591e-05, |
|
"loss": 0.3636, |
|
"step": 236000 |
|
}, |
|
{ |
|
"epoch": 10.53, |
|
"eval_loss": 0.3383854627609253, |
|
"eval_runtime": 2.1919, |
|
"eval_samples_per_second": 1047.95, |
|
"eval_steps_per_second": 16.424, |
|
"step": 236000 |
|
}, |
|
{ |
|
"epoch": 10.55, |
|
"learning_rate": 1.4691132166287069e-05, |
|
"loss": 0.3638, |
|
"step": 236500 |
|
}, |
|
{ |
|
"epoch": 10.57, |
|
"learning_rate": 1.4350916865007609e-05, |
|
"loss": 0.3636, |
|
"step": 237000 |
|
}, |
|
{ |
|
"epoch": 10.57, |
|
"eval_loss": 0.33673885464668274, |
|
"eval_runtime": 2.1228, |
|
"eval_samples_per_second": 1082.067, |
|
"eval_steps_per_second": 16.959, |
|
"step": 237000 |
|
}, |
|
{ |
|
"epoch": 10.6, |
|
"learning_rate": 1.4023415496196915e-05, |
|
"loss": 0.3635, |
|
"step": 237500 |
|
}, |
|
{ |
|
"epoch": 10.62, |
|
"learning_rate": 1.3708642385825806e-05, |
|
"loss": 0.3638, |
|
"step": 238000 |
|
}, |
|
{ |
|
"epoch": 10.62, |
|
"eval_loss": 0.3373904228210449, |
|
"eval_runtime": 2.309, |
|
"eval_samples_per_second": 994.807, |
|
"eval_steps_per_second": 15.591, |
|
"step": 238000 |
|
}, |
|
{ |
|
"epoch": 10.64, |
|
"learning_rate": 1.3406611303089918e-05, |
|
"loss": 0.3639, |
|
"step": 238500 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 1.311733545980718e-05, |
|
"loss": 0.3637, |
|
"step": 239000 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"eval_loss": 0.3368372619152069, |
|
"eval_runtime": 2.2218, |
|
"eval_samples_per_second": 1033.83, |
|
"eval_steps_per_second": 16.203, |
|
"step": 239000 |
|
}, |
|
{ |
|
"epoch": 10.68, |
|
"learning_rate": 1.284082750984018e-05, |
|
"loss": 0.3635, |
|
"step": 239500 |
|
}, |
|
{ |
|
"epoch": 10.71, |
|
"learning_rate": 1.2577099548542319e-05, |
|
"loss": 0.3635, |
|
"step": 240000 |
|
}, |
|
{ |
|
"epoch": 10.71, |
|
"eval_loss": 0.33518004417419434, |
|
"eval_runtime": 2.1947, |
|
"eval_samples_per_second": 1046.621, |
|
"eval_steps_per_second": 16.403, |
|
"step": 240000 |
|
}, |
|
{ |
|
"epoch": 10.73, |
|
"learning_rate": 1.232616311222906e-05, |
|
"loss": 0.3635, |
|
"step": 240500 |
|
}, |
|
{ |
|
"epoch": 10.75, |
|
"learning_rate": 1.2088029177672978e-05, |
|
"loss": 0.3635, |
|
"step": 241000 |
|
}, |
|
{ |
|
"epoch": 10.75, |
|
"eval_loss": 0.3392580449581146, |
|
"eval_runtime": 2.2637, |
|
"eval_samples_per_second": 1014.699, |
|
"eval_steps_per_second": 15.903, |
|
"step": 241000 |
|
}, |
|
{ |
|
"epoch": 10.77, |
|
"learning_rate": 1.1862708161623889e-05, |
|
"loss": 0.3634, |
|
"step": 241500 |
|
}, |
|
{ |
|
"epoch": 10.8, |
|
"learning_rate": 1.165020992035296e-05, |
|
"loss": 0.3634, |
|
"step": 242000 |
|
}, |
|
{ |
|
"epoch": 10.8, |
|
"eval_loss": 0.3343862295150757, |
|
"eval_runtime": 2.2213, |
|
"eval_samples_per_second": 1034.066, |
|
"eval_steps_per_second": 16.207, |
|
"step": 242000 |
|
}, |
|
{ |
|
"epoch": 10.82, |
|
"learning_rate": 1.1450543749221632e-05, |
|
"loss": 0.3635, |
|
"step": 242500 |
|
}, |
|
{ |
|
"epoch": 10.84, |
|
"learning_rate": 1.126371838227509e-05, |
|
"loss": 0.3635, |
|
"step": 243000 |
|
}, |
|
{ |
|
"epoch": 10.84, |
|
"eval_loss": 0.3382641077041626, |
|
"eval_runtime": 2.2515, |
|
"eval_samples_per_second": 1020.218, |
|
"eval_steps_per_second": 15.989, |
|
"step": 243000 |
|
}, |
|
{ |
|
"epoch": 10.86, |
|
"learning_rate": 1.1089741991860081e-05, |
|
"loss": 0.3634, |
|
"step": 243500 |
|
}, |
|
{ |
|
"epoch": 10.89, |
|
"learning_rate": 1.0928622188267536e-05, |
|
"loss": 0.3633, |
|
"step": 244000 |
|
}, |
|
{ |
|
"epoch": 10.89, |
|
"eval_loss": 0.33619487285614014, |
|
"eval_runtime": 2.2703, |
|
"eval_samples_per_second": 1011.748, |
|
"eval_steps_per_second": 15.857, |
|
"step": 244000 |
|
}, |
|
{ |
|
"epoch": 10.91, |
|
"learning_rate": 1.0780366019399665e-05, |
|
"loss": 0.3635, |
|
"step": 244500 |
|
}, |
|
{ |
|
"epoch": 10.93, |
|
"learning_rate": 1.0644979970461512e-05, |
|
"loss": 0.3635, |
|
"step": 245000 |
|
}, |
|
{ |
|
"epoch": 10.93, |
|
"eval_loss": 0.3352552354335785, |
|
"eval_runtime": 2.2392, |
|
"eval_samples_per_second": 1025.803, |
|
"eval_steps_per_second": 16.077, |
|
"step": 245000 |
|
}, |
|
{ |
|
"epoch": 10.95, |
|
"learning_rate": 1.0522469963677483e-05, |
|
"loss": 0.3634, |
|
"step": 245500 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"learning_rate": 1.0412841358032126e-05, |
|
"loss": 0.3634, |
|
"step": 246000 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"eval_loss": 0.33572423458099365, |
|
"eval_runtime": 2.2818, |
|
"eval_samples_per_second": 1006.659, |
|
"eval_steps_per_second": 15.777, |
|
"step": 246000 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 1.0316098949035847e-05, |
|
"loss": 0.3634, |
|
"step": 246500 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 1.0232246968514984e-05, |
|
"loss": 0.3632, |
|
"step": 247000 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"eval_loss": 0.3374698758125305, |
|
"eval_runtime": 3.6783, |
|
"eval_samples_per_second": 624.479, |
|
"eval_steps_per_second": 9.787, |
|
"step": 247000 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 1.0161289084426815e-05, |
|
"loss": 0.3633, |
|
"step": 247500 |
|
}, |
|
{ |
|
"epoch": 11.06, |
|
"learning_rate": 1.0103228400699063e-05, |
|
"loss": 0.3633, |
|
"step": 248000 |
|
}, |
|
{ |
|
"epoch": 11.06, |
|
"eval_loss": 0.3395208418369293, |
|
"eval_runtime": 2.2423, |
|
"eval_samples_per_second": 1024.379, |
|
"eval_steps_per_second": 16.055, |
|
"step": 248000 |
|
}, |
|
{ |
|
"epoch": 11.09, |
|
"learning_rate": 1.0058067457094136e-05, |
|
"loss": 0.3635, |
|
"step": 248500 |
|
}, |
|
{ |
|
"epoch": 11.11, |
|
"learning_rate": 1.0025808229097982e-05, |
|
"loss": 0.3635, |
|
"step": 249000 |
|
}, |
|
{ |
|
"epoch": 11.11, |
|
"eval_loss": 0.33821383118629456, |
|
"eval_runtime": 2.2205, |
|
"eval_samples_per_second": 1034.472, |
|
"eval_steps_per_second": 16.213, |
|
"step": 249000 |
|
}, |
|
{ |
|
"epoch": 11.13, |
|
"learning_rate": 1.0006452127833747e-05, |
|
"loss": 0.3633, |
|
"step": 249500 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"learning_rate": 1e-05, |
|
"loss": 0.3634, |
|
"step": 250000 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"eval_loss": 0.3380183279514313, |
|
"eval_runtime": 2.1982, |
|
"eval_samples_per_second": 1044.959, |
|
"eval_steps_per_second": 16.377, |
|
"step": 250000 |
|
} |
|
], |
|
"max_steps": 250000, |
|
"num_train_epochs": 12, |
|
"total_flos": 4.004121619841193e+21, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|