|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.29000429000429, |
|
"eval_steps": 1000, |
|
"global_step": 5000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02145002145002145, |
|
"grad_norm": 9.066595077514648, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 1.715, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0429000429000429, |
|
"grad_norm": 6.495127201080322, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 1.3703, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.06435006435006435, |
|
"grad_norm": 5.474078178405762, |
|
"learning_rate": 1.5e-06, |
|
"loss": 1.0103, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.0858000858000858, |
|
"grad_norm": 5.429323196411133, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.9033, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.10725010725010725, |
|
"grad_norm": 5.380603790283203, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.7779, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.1287001287001287, |
|
"grad_norm": 4.108312606811523, |
|
"learning_rate": 3e-06, |
|
"loss": 0.7064, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.15015015015015015, |
|
"grad_norm": 4.261275291442871, |
|
"learning_rate": 3.5e-06, |
|
"loss": 0.6859, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.1716001716001716, |
|
"grad_norm": 4.792464256286621, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.6422, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.19305019305019305, |
|
"grad_norm": 4.474643707275391, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.6719, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.2145002145002145, |
|
"grad_norm": 5.082165718078613, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6536, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.23595023595023595, |
|
"grad_norm": 5.100755214691162, |
|
"learning_rate": 5.500000000000001e-06, |
|
"loss": 0.6508, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.2574002574002574, |
|
"grad_norm": 4.228156566619873, |
|
"learning_rate": 6e-06, |
|
"loss": 0.633, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.27885027885027885, |
|
"grad_norm": 4.423853874206543, |
|
"learning_rate": 6.5000000000000004e-06, |
|
"loss": 0.5827, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.3003003003003003, |
|
"grad_norm": 4.693906784057617, |
|
"learning_rate": 7e-06, |
|
"loss": 0.5881, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.32175032175032175, |
|
"grad_norm": 5.522333145141602, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.5953, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.3432003432003432, |
|
"grad_norm": 3.7150588035583496, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.5733, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.36465036465036466, |
|
"grad_norm": 4.65098762512207, |
|
"learning_rate": 8.5e-06, |
|
"loss": 0.5741, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.3861003861003861, |
|
"grad_norm": 4.152739524841309, |
|
"learning_rate": 9e-06, |
|
"loss": 0.5721, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.40755040755040756, |
|
"grad_norm": 4.5913615226745605, |
|
"learning_rate": 9.5e-06, |
|
"loss": 0.5837, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.429000429000429, |
|
"grad_norm": 3.7452423572540283, |
|
"learning_rate": 1e-05, |
|
"loss": 0.5763, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.45045045045045046, |
|
"grad_norm": 4.0456953048706055, |
|
"learning_rate": 9.944444444444445e-06, |
|
"loss": 0.5743, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.4719004719004719, |
|
"grad_norm": 4.183647155761719, |
|
"learning_rate": 9.88888888888889e-06, |
|
"loss": 0.5851, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.49335049335049336, |
|
"grad_norm": 3.6905064582824707, |
|
"learning_rate": 9.833333333333333e-06, |
|
"loss": 0.5423, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.5148005148005148, |
|
"grad_norm": 4.111310005187988, |
|
"learning_rate": 9.777777777777779e-06, |
|
"loss": 0.5392, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5362505362505362, |
|
"grad_norm": 3.7859528064727783, |
|
"learning_rate": 9.722222222222223e-06, |
|
"loss": 0.5257, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.5577005577005577, |
|
"grad_norm": 3.810257911682129, |
|
"learning_rate": 9.666666666666667e-06, |
|
"loss": 0.4958, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.5791505791505791, |
|
"grad_norm": 3.279156446456909, |
|
"learning_rate": 9.611111111111112e-06, |
|
"loss": 0.5103, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.6006006006006006, |
|
"grad_norm": 3.16200852394104, |
|
"learning_rate": 9.555555555555556e-06, |
|
"loss": 0.5262, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.622050622050622, |
|
"grad_norm": 3.9035494327545166, |
|
"learning_rate": 9.5e-06, |
|
"loss": 0.5131, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.6435006435006435, |
|
"grad_norm": 3.071242570877075, |
|
"learning_rate": 9.444444444444445e-06, |
|
"loss": 0.478, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.6649506649506649, |
|
"grad_norm": 4.14898157119751, |
|
"learning_rate": 9.38888888888889e-06, |
|
"loss": 0.5035, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.6864006864006864, |
|
"grad_norm": 3.8921945095062256, |
|
"learning_rate": 9.333333333333334e-06, |
|
"loss": 0.5, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7078507078507078, |
|
"grad_norm": 3.4418222904205322, |
|
"learning_rate": 9.277777777777778e-06, |
|
"loss": 0.4676, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.7293007293007293, |
|
"grad_norm": 3.1406946182250977, |
|
"learning_rate": 9.222222222222224e-06, |
|
"loss": 0.4668, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.7507507507507507, |
|
"grad_norm": 3.900789499282837, |
|
"learning_rate": 9.166666666666666e-06, |
|
"loss": 0.4834, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.7722007722007722, |
|
"grad_norm": 3.2185850143432617, |
|
"learning_rate": 9.111111111111112e-06, |
|
"loss": 0.4716, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7936507936507936, |
|
"grad_norm": 3.683173179626465, |
|
"learning_rate": 9.055555555555556e-06, |
|
"loss": 0.4882, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.8151008151008151, |
|
"grad_norm": 3.3974883556365967, |
|
"learning_rate": 9e-06, |
|
"loss": 0.4781, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.8365508365508365, |
|
"grad_norm": 3.78350830078125, |
|
"learning_rate": 8.944444444444446e-06, |
|
"loss": 0.4872, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.858000858000858, |
|
"grad_norm": 3.633676528930664, |
|
"learning_rate": 8.888888888888888e-06, |
|
"loss": 0.4429, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.858000858000858, |
|
"eval_loss": 0.46728599071502686, |
|
"eval_runtime": 1674.249, |
|
"eval_samples_per_second": 2.33, |
|
"eval_steps_per_second": 0.146, |
|
"eval_wer": 0.3495363662706462, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8794508794508794, |
|
"grad_norm": 3.6749162673950195, |
|
"learning_rate": 8.833333333333334e-06, |
|
"loss": 0.4712, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.9009009009009009, |
|
"grad_norm": 3.0788767337799072, |
|
"learning_rate": 8.777777777777778e-06, |
|
"loss": 0.4857, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.9223509223509223, |
|
"grad_norm": 3.3249752521514893, |
|
"learning_rate": 8.722222222222224e-06, |
|
"loss": 0.4406, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.9438009438009438, |
|
"grad_norm": 3.189654588699341, |
|
"learning_rate": 8.666666666666668e-06, |
|
"loss": 0.4313, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9652509652509652, |
|
"grad_norm": 3.3606882095336914, |
|
"learning_rate": 8.611111111111112e-06, |
|
"loss": 0.4517, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.9867009867009867, |
|
"grad_norm": 3.4091646671295166, |
|
"learning_rate": 8.555555555555556e-06, |
|
"loss": 0.4378, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.008151008151008, |
|
"grad_norm": 2.8111917972564697, |
|
"learning_rate": 8.5e-06, |
|
"loss": 0.3944, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 1.0296010296010296, |
|
"grad_norm": 2.7863500118255615, |
|
"learning_rate": 8.444444444444446e-06, |
|
"loss": 0.3242, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.0510510510510511, |
|
"grad_norm": 2.681203603744507, |
|
"learning_rate": 8.38888888888889e-06, |
|
"loss": 0.3294, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 1.0725010725010724, |
|
"grad_norm": 3.21625018119812, |
|
"learning_rate": 8.333333333333334e-06, |
|
"loss": 0.3232, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.093951093951094, |
|
"grad_norm": 2.1063191890716553, |
|
"learning_rate": 8.277777777777778e-06, |
|
"loss": 0.3162, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 1.1154011154011154, |
|
"grad_norm": 2.7239444255828857, |
|
"learning_rate": 8.222222222222222e-06, |
|
"loss": 0.3372, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.136851136851137, |
|
"grad_norm": 2.6881861686706543, |
|
"learning_rate": 8.166666666666668e-06, |
|
"loss": 0.3269, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 1.1583011583011582, |
|
"grad_norm": 3.041693925857544, |
|
"learning_rate": 8.111111111111112e-06, |
|
"loss": 0.3299, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.1797511797511797, |
|
"grad_norm": 3.6576833724975586, |
|
"learning_rate": 8.055555555555557e-06, |
|
"loss": 0.3786, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 1.2012012012012012, |
|
"grad_norm": 3.035702705383301, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.346, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.2226512226512227, |
|
"grad_norm": 2.5944619178771973, |
|
"learning_rate": 7.944444444444445e-06, |
|
"loss": 0.3309, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.244101244101244, |
|
"grad_norm": 2.5746264457702637, |
|
"learning_rate": 7.88888888888889e-06, |
|
"loss": 0.3235, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.2655512655512655, |
|
"grad_norm": 2.7788243293762207, |
|
"learning_rate": 7.833333333333333e-06, |
|
"loss": 0.3212, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.287001287001287, |
|
"grad_norm": 3.6462416648864746, |
|
"learning_rate": 7.77777777777778e-06, |
|
"loss": 0.332, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.3084513084513085, |
|
"grad_norm": 2.2463419437408447, |
|
"learning_rate": 7.722222222222223e-06, |
|
"loss": 0.3366, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 1.3299013299013298, |
|
"grad_norm": 3.5368576049804688, |
|
"learning_rate": 7.666666666666667e-06, |
|
"loss": 0.3138, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.3513513513513513, |
|
"grad_norm": 2.547539472579956, |
|
"learning_rate": 7.611111111111111e-06, |
|
"loss": 0.3564, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.3728013728013728, |
|
"grad_norm": 3.0548551082611084, |
|
"learning_rate": 7.555555555555556e-06, |
|
"loss": 0.3393, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.3942513942513943, |
|
"grad_norm": 2.7717063426971436, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.3266, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 1.4157014157014158, |
|
"grad_norm": 2.6384313106536865, |
|
"learning_rate": 7.444444444444445e-06, |
|
"loss": 0.3355, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.4371514371514371, |
|
"grad_norm": 2.7044169902801514, |
|
"learning_rate": 7.38888888888889e-06, |
|
"loss": 0.3375, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 1.4586014586014586, |
|
"grad_norm": 2.4753165245056152, |
|
"learning_rate": 7.333333333333333e-06, |
|
"loss": 0.3347, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.4800514800514801, |
|
"grad_norm": 3.0648443698883057, |
|
"learning_rate": 7.277777777777778e-06, |
|
"loss": 0.3202, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.5015015015015014, |
|
"grad_norm": 3.1359264850616455, |
|
"learning_rate": 7.222222222222223e-06, |
|
"loss": 0.3303, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.522951522951523, |
|
"grad_norm": 2.995232105255127, |
|
"learning_rate": 7.166666666666667e-06, |
|
"loss": 0.3197, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 1.5444015444015444, |
|
"grad_norm": 2.7742152214050293, |
|
"learning_rate": 7.111111111111112e-06, |
|
"loss": 0.3255, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.565851565851566, |
|
"grad_norm": 2.7029542922973633, |
|
"learning_rate": 7.055555555555557e-06, |
|
"loss": 0.3147, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 1.5873015873015874, |
|
"grad_norm": 2.604520082473755, |
|
"learning_rate": 7e-06, |
|
"loss": 0.3235, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.6087516087516087, |
|
"grad_norm": 3.152074098587036, |
|
"learning_rate": 6.944444444444445e-06, |
|
"loss": 0.319, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 1.6302016302016302, |
|
"grad_norm": 3.121864080429077, |
|
"learning_rate": 6.88888888888889e-06, |
|
"loss": 0.2972, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.6516516516516515, |
|
"grad_norm": 2.733600616455078, |
|
"learning_rate": 6.833333333333334e-06, |
|
"loss": 0.3177, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 1.673101673101673, |
|
"grad_norm": 2.8984882831573486, |
|
"learning_rate": 6.777777777777779e-06, |
|
"loss": 0.3142, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.6945516945516945, |
|
"grad_norm": 2.585207939147949, |
|
"learning_rate": 6.7222222222222235e-06, |
|
"loss": 0.3372, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 1.716001716001716, |
|
"grad_norm": 3.007678985595703, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 0.3192, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.716001716001716, |
|
"eval_loss": 0.41164642572402954, |
|
"eval_runtime": 1704.3317, |
|
"eval_samples_per_second": 2.289, |
|
"eval_steps_per_second": 0.143, |
|
"eval_wer": 0.2986090988119386, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.7374517374517375, |
|
"grad_norm": 2.6128337383270264, |
|
"learning_rate": 6.6111111111111115e-06, |
|
"loss": 0.3215, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 1.758901758901759, |
|
"grad_norm": 2.6473169326782227, |
|
"learning_rate": 6.555555555555556e-06, |
|
"loss": 0.324, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.7803517803517803, |
|
"grad_norm": 2.654451608657837, |
|
"learning_rate": 6.5000000000000004e-06, |
|
"loss": 0.3029, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 1.8018018018018018, |
|
"grad_norm": 2.335432767868042, |
|
"learning_rate": 6.444444444444445e-06, |
|
"loss": 0.3197, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.8232518232518231, |
|
"grad_norm": 2.4517362117767334, |
|
"learning_rate": 6.3888888888888885e-06, |
|
"loss": 0.3136, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 1.8447018447018446, |
|
"grad_norm": 2.7718915939331055, |
|
"learning_rate": 6.333333333333333e-06, |
|
"loss": 0.3059, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.8661518661518661, |
|
"grad_norm": 2.876389265060425, |
|
"learning_rate": 6.277777777777778e-06, |
|
"loss": 0.3193, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 1.8876018876018876, |
|
"grad_norm": 2.892172336578369, |
|
"learning_rate": 6.222222222222223e-06, |
|
"loss": 0.3097, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.9090519090519091, |
|
"grad_norm": 3.030829429626465, |
|
"learning_rate": 6.166666666666667e-06, |
|
"loss": 0.3232, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 1.9305019305019306, |
|
"grad_norm": 2.721433162689209, |
|
"learning_rate": 6.111111111111112e-06, |
|
"loss": 0.2944, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.951951951951952, |
|
"grad_norm": 3.1279942989349365, |
|
"learning_rate": 6.055555555555555e-06, |
|
"loss": 0.3063, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 1.9734019734019734, |
|
"grad_norm": 2.706367254257202, |
|
"learning_rate": 6e-06, |
|
"loss": 0.2998, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.9948519948519947, |
|
"grad_norm": 3.0728838443756104, |
|
"learning_rate": 5.944444444444445e-06, |
|
"loss": 0.3036, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 2.016302016302016, |
|
"grad_norm": 2.0992302894592285, |
|
"learning_rate": 5.88888888888889e-06, |
|
"loss": 0.2477, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.0377520377520377, |
|
"grad_norm": 1.931810975074768, |
|
"learning_rate": 5.833333333333334e-06, |
|
"loss": 0.2085, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 2.0592020592020592, |
|
"grad_norm": 2.4714324474334717, |
|
"learning_rate": 5.777777777777778e-06, |
|
"loss": 0.201, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.0806520806520807, |
|
"grad_norm": 2.1515088081359863, |
|
"learning_rate": 5.722222222222222e-06, |
|
"loss": 0.1978, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 2.1021021021021022, |
|
"grad_norm": 2.203587770462036, |
|
"learning_rate": 5.666666666666667e-06, |
|
"loss": 0.2082, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 2.1235521235521237, |
|
"grad_norm": 2.3945024013519287, |
|
"learning_rate": 5.611111111111112e-06, |
|
"loss": 0.2012, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 2.145002145002145, |
|
"grad_norm": 2.5214550495147705, |
|
"learning_rate": 5.555555555555557e-06, |
|
"loss": 0.1983, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.1664521664521663, |
|
"grad_norm": 2.8749985694885254, |
|
"learning_rate": 5.500000000000001e-06, |
|
"loss": 0.202, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 2.187902187902188, |
|
"grad_norm": 2.302109956741333, |
|
"learning_rate": 5.444444444444445e-06, |
|
"loss": 0.187, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.2093522093522093, |
|
"grad_norm": 2.303842782974243, |
|
"learning_rate": 5.388888888888889e-06, |
|
"loss": 0.2007, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 2.230802230802231, |
|
"grad_norm": 2.1499674320220947, |
|
"learning_rate": 5.333333333333334e-06, |
|
"loss": 0.2025, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.2522522522522523, |
|
"grad_norm": 2.735565185546875, |
|
"learning_rate": 5.2777777777777785e-06, |
|
"loss": 0.1934, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 2.273702273702274, |
|
"grad_norm": 2.781230926513672, |
|
"learning_rate": 5.2222222222222226e-06, |
|
"loss": 0.1959, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.2951522951522954, |
|
"grad_norm": 2.934748649597168, |
|
"learning_rate": 5.1666666666666675e-06, |
|
"loss": 0.1923, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 2.3166023166023164, |
|
"grad_norm": 2.5688586235046387, |
|
"learning_rate": 5.1111111111111115e-06, |
|
"loss": 0.2096, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.338052338052338, |
|
"grad_norm": 2.15342378616333, |
|
"learning_rate": 5.0555555555555555e-06, |
|
"loss": 0.198, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 2.3595023595023594, |
|
"grad_norm": 2.287332773208618, |
|
"learning_rate": 5e-06, |
|
"loss": 0.2079, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.380952380952381, |
|
"grad_norm": 2.2606818675994873, |
|
"learning_rate": 4.944444444444445e-06, |
|
"loss": 0.2014, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 2.4024024024024024, |
|
"grad_norm": 2.0348024368286133, |
|
"learning_rate": 4.888888888888889e-06, |
|
"loss": 0.2046, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.423852423852424, |
|
"grad_norm": 2.326032876968384, |
|
"learning_rate": 4.833333333333333e-06, |
|
"loss": 0.1905, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 2.4453024453024454, |
|
"grad_norm": 2.2641820907592773, |
|
"learning_rate": 4.777777777777778e-06, |
|
"loss": 0.1959, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.4667524667524665, |
|
"grad_norm": 1.923122763633728, |
|
"learning_rate": 4.722222222222222e-06, |
|
"loss": 0.199, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 2.488202488202488, |
|
"grad_norm": 2.273163080215454, |
|
"learning_rate": 4.666666666666667e-06, |
|
"loss": 0.1983, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.5096525096525095, |
|
"grad_norm": 2.816136598587036, |
|
"learning_rate": 4.611111111111112e-06, |
|
"loss": 0.1901, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 2.531102531102531, |
|
"grad_norm": 2.453509569168091, |
|
"learning_rate": 4.555555555555556e-06, |
|
"loss": 0.1819, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.5525525525525525, |
|
"grad_norm": 1.754265546798706, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.1808, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 2.574002574002574, |
|
"grad_norm": 2.362611770629883, |
|
"learning_rate": 4.444444444444444e-06, |
|
"loss": 0.1917, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.574002574002574, |
|
"eval_loss": 0.4086485207080841, |
|
"eval_runtime": 1612.1909, |
|
"eval_samples_per_second": 2.42, |
|
"eval_steps_per_second": 0.151, |
|
"eval_wer": 0.2937071380276248, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.5954525954525955, |
|
"grad_norm": 2.633920669555664, |
|
"learning_rate": 4.388888888888889e-06, |
|
"loss": 0.188, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 2.616902616902617, |
|
"grad_norm": 2.290724515914917, |
|
"learning_rate": 4.333333333333334e-06, |
|
"loss": 0.2027, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.6383526383526386, |
|
"grad_norm": 2.451239585876465, |
|
"learning_rate": 4.277777777777778e-06, |
|
"loss": 0.1982, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 2.6598026598026596, |
|
"grad_norm": 2.550907611846924, |
|
"learning_rate": 4.222222222222223e-06, |
|
"loss": 0.1994, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.681252681252681, |
|
"grad_norm": 2.633718729019165, |
|
"learning_rate": 4.166666666666667e-06, |
|
"loss": 0.229, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 2.7027027027027026, |
|
"grad_norm": 2.618665933609009, |
|
"learning_rate": 4.111111111111111e-06, |
|
"loss": 0.2017, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.724152724152724, |
|
"grad_norm": 2.954728364944458, |
|
"learning_rate": 4.055555555555556e-06, |
|
"loss": 0.1873, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 2.7456027456027456, |
|
"grad_norm": 2.27900767326355, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.2034, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.767052767052767, |
|
"grad_norm": 2.202578544616699, |
|
"learning_rate": 3.944444444444445e-06, |
|
"loss": 0.1955, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 2.7885027885027887, |
|
"grad_norm": 2.391146659851074, |
|
"learning_rate": 3.88888888888889e-06, |
|
"loss": 0.1934, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.8099528099528097, |
|
"grad_norm": 2.5382649898529053, |
|
"learning_rate": 3.833333333333334e-06, |
|
"loss": 0.1907, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 2.8314028314028317, |
|
"grad_norm": 2.9581174850463867, |
|
"learning_rate": 3.777777777777778e-06, |
|
"loss": 0.2083, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.8528528528528527, |
|
"grad_norm": 2.907230854034424, |
|
"learning_rate": 3.7222222222222225e-06, |
|
"loss": 0.2074, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 2.8743028743028742, |
|
"grad_norm": 2.437242031097412, |
|
"learning_rate": 3.6666666666666666e-06, |
|
"loss": 0.193, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 2.8957528957528957, |
|
"grad_norm": 2.3380300998687744, |
|
"learning_rate": 3.6111111111111115e-06, |
|
"loss": 0.211, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 2.9172029172029172, |
|
"grad_norm": 3.0222008228302, |
|
"learning_rate": 3.555555555555556e-06, |
|
"loss": 0.1924, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.9386529386529388, |
|
"grad_norm": 2.763075351715088, |
|
"learning_rate": 3.5e-06, |
|
"loss": 0.1989, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 2.9601029601029603, |
|
"grad_norm": 2.859872817993164, |
|
"learning_rate": 3.444444444444445e-06, |
|
"loss": 0.1845, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 2.9815529815529818, |
|
"grad_norm": 2.663336992263794, |
|
"learning_rate": 3.3888888888888893e-06, |
|
"loss": 0.1994, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 3.003003003003003, |
|
"grad_norm": 1.762622594833374, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 0.1894, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 3.0244530244530243, |
|
"grad_norm": 1.7877063751220703, |
|
"learning_rate": 3.277777777777778e-06, |
|
"loss": 0.1154, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 3.045903045903046, |
|
"grad_norm": 2.131319761276245, |
|
"learning_rate": 3.2222222222222227e-06, |
|
"loss": 0.1122, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 3.0673530673530673, |
|
"grad_norm": 1.935207486152649, |
|
"learning_rate": 3.1666666666666667e-06, |
|
"loss": 0.1157, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 3.088803088803089, |
|
"grad_norm": 2.1667850017547607, |
|
"learning_rate": 3.1111111111111116e-06, |
|
"loss": 0.1165, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 3.1102531102531104, |
|
"grad_norm": 2.0937955379486084, |
|
"learning_rate": 3.055555555555556e-06, |
|
"loss": 0.1066, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 3.131703131703132, |
|
"grad_norm": 2.4619686603546143, |
|
"learning_rate": 3e-06, |
|
"loss": 0.1047, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 3.153153153153153, |
|
"grad_norm": 2.3083388805389404, |
|
"learning_rate": 2.944444444444445e-06, |
|
"loss": 0.1483, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 3.1746031746031744, |
|
"grad_norm": 2.0462327003479004, |
|
"learning_rate": 2.888888888888889e-06, |
|
"loss": 0.1183, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 3.196053196053196, |
|
"grad_norm": 2.2849583625793457, |
|
"learning_rate": 2.8333333333333335e-06, |
|
"loss": 0.1155, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 3.2175032175032174, |
|
"grad_norm": 2.364773750305176, |
|
"learning_rate": 2.7777777777777783e-06, |
|
"loss": 0.1139, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 3.238953238953239, |
|
"grad_norm": 2.6987013816833496, |
|
"learning_rate": 2.7222222222222224e-06, |
|
"loss": 0.1117, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 3.2604032604032605, |
|
"grad_norm": 2.8335797786712646, |
|
"learning_rate": 2.666666666666667e-06, |
|
"loss": 0.1135, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 3.281853281853282, |
|
"grad_norm": 2.0935230255126953, |
|
"learning_rate": 2.6111111111111113e-06, |
|
"loss": 0.1048, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 3.3033033033033035, |
|
"grad_norm": 2.1639814376831055, |
|
"learning_rate": 2.5555555555555557e-06, |
|
"loss": 0.1088, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 3.324753324753325, |
|
"grad_norm": 2.022599458694458, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.1093, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 3.346203346203346, |
|
"grad_norm": 2.6962931156158447, |
|
"learning_rate": 2.4444444444444447e-06, |
|
"loss": 0.1139, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 3.3676533676533675, |
|
"grad_norm": 1.9098910093307495, |
|
"learning_rate": 2.388888888888889e-06, |
|
"loss": 0.1082, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 3.389103389103389, |
|
"grad_norm": 2.3071112632751465, |
|
"learning_rate": 2.3333333333333336e-06, |
|
"loss": 0.129, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 3.4105534105534105, |
|
"grad_norm": 1.9279221296310425, |
|
"learning_rate": 2.277777777777778e-06, |
|
"loss": 0.1153, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 3.432003432003432, |
|
"grad_norm": 2.1972861289978027, |
|
"learning_rate": 2.222222222222222e-06, |
|
"loss": 0.1113, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.432003432003432, |
|
"eval_loss": 0.43410712480545044, |
|
"eval_runtime": 2229.8693, |
|
"eval_samples_per_second": 1.749, |
|
"eval_steps_per_second": 0.109, |
|
"eval_wer": 0.2852313339128755, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.4534534534534536, |
|
"grad_norm": 1.9216057062149048, |
|
"learning_rate": 2.166666666666667e-06, |
|
"loss": 0.1126, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 3.474903474903475, |
|
"grad_norm": 2.277034282684326, |
|
"learning_rate": 2.1111111111111114e-06, |
|
"loss": 0.1088, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 3.496353496353496, |
|
"grad_norm": 2.0428617000579834, |
|
"learning_rate": 2.0555555555555555e-06, |
|
"loss": 0.1083, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 3.517803517803518, |
|
"grad_norm": 2.5672316551208496, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.1059, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 3.539253539253539, |
|
"grad_norm": 2.100874185562134, |
|
"learning_rate": 1.944444444444445e-06, |
|
"loss": 0.1118, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 3.5607035607035606, |
|
"grad_norm": 2.066019296646118, |
|
"learning_rate": 1.888888888888889e-06, |
|
"loss": 0.1043, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 3.582153582153582, |
|
"grad_norm": 2.2962706089019775, |
|
"learning_rate": 1.8333333333333333e-06, |
|
"loss": 0.1158, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 3.6036036036036037, |
|
"grad_norm": 2.4337549209594727, |
|
"learning_rate": 1.777777777777778e-06, |
|
"loss": 0.1074, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.625053625053625, |
|
"grad_norm": 2.21247935295105, |
|
"learning_rate": 1.7222222222222224e-06, |
|
"loss": 0.1081, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 3.6465036465036467, |
|
"grad_norm": 1.9160076379776, |
|
"learning_rate": 1.6666666666666667e-06, |
|
"loss": 0.1139, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 3.667953667953668, |
|
"grad_norm": 2.366440773010254, |
|
"learning_rate": 1.6111111111111113e-06, |
|
"loss": 0.1072, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 3.6894036894036892, |
|
"grad_norm": 2.299656867980957, |
|
"learning_rate": 1.5555555555555558e-06, |
|
"loss": 0.1108, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.7108537108537107, |
|
"grad_norm": 2.0302138328552246, |
|
"learning_rate": 1.5e-06, |
|
"loss": 0.1013, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 3.7323037323037322, |
|
"grad_norm": 2.416654586791992, |
|
"learning_rate": 1.4444444444444445e-06, |
|
"loss": 0.1134, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 3.7537537537537538, |
|
"grad_norm": 1.7967655658721924, |
|
"learning_rate": 1.3888888888888892e-06, |
|
"loss": 0.1001, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 3.7752037752037753, |
|
"grad_norm": 2.470515251159668, |
|
"learning_rate": 1.3333333333333334e-06, |
|
"loss": 0.1033, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.7966537966537968, |
|
"grad_norm": 2.558213949203491, |
|
"learning_rate": 1.2777777777777779e-06, |
|
"loss": 0.1081, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 3.8181038181038183, |
|
"grad_norm": 2.0769007205963135, |
|
"learning_rate": 1.2222222222222223e-06, |
|
"loss": 0.1027, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 3.8395538395538393, |
|
"grad_norm": 1.9774547815322876, |
|
"learning_rate": 1.1666666666666668e-06, |
|
"loss": 0.1136, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 3.861003861003861, |
|
"grad_norm": 2.82588791847229, |
|
"learning_rate": 1.111111111111111e-06, |
|
"loss": 0.1129, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.8824538824538823, |
|
"grad_norm": 2.3755438327789307, |
|
"learning_rate": 1.0555555555555557e-06, |
|
"loss": 0.1095, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 3.903903903903904, |
|
"grad_norm": 2.441087007522583, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 0.1007, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 3.9253539253539254, |
|
"grad_norm": 2.604708433151245, |
|
"learning_rate": 9.444444444444445e-07, |
|
"loss": 0.1055, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 3.946803946803947, |
|
"grad_norm": 2.5220184326171875, |
|
"learning_rate": 8.88888888888889e-07, |
|
"loss": 0.112, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.9682539682539684, |
|
"grad_norm": 2.3847906589508057, |
|
"learning_rate": 8.333333333333333e-07, |
|
"loss": 0.1082, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 3.9897039897039894, |
|
"grad_norm": 1.749206304550171, |
|
"learning_rate": 7.8e-07, |
|
"loss": 0.1096, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 4.011154011154011, |
|
"grad_norm": 1.657009482383728, |
|
"learning_rate": 7.244444444444446e-07, |
|
"loss": 0.084, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 4.032604032604032, |
|
"grad_norm": 1.4502904415130615, |
|
"learning_rate": 6.68888888888889e-07, |
|
"loss": 0.0624, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 4.054054054054054, |
|
"grad_norm": 1.753985047340393, |
|
"learning_rate": 6.133333333333333e-07, |
|
"loss": 0.064, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 4.0755040755040755, |
|
"grad_norm": 1.8382900953292847, |
|
"learning_rate": 5.577777777777779e-07, |
|
"loss": 0.0607, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 4.0969540969540965, |
|
"grad_norm": 2.292609930038452, |
|
"learning_rate": 5.022222222222222e-07, |
|
"loss": 0.0617, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 4.1184041184041185, |
|
"grad_norm": 1.5938409566879272, |
|
"learning_rate": 4.466666666666667e-07, |
|
"loss": 0.0643, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 4.1398541398541395, |
|
"grad_norm": 1.7310892343521118, |
|
"learning_rate": 3.9111111111111115e-07, |
|
"loss": 0.0595, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 4.1613041613041615, |
|
"grad_norm": 1.9981136322021484, |
|
"learning_rate": 3.3555555555555556e-07, |
|
"loss": 0.0553, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 4.1827541827541825, |
|
"grad_norm": 1.4661279916763306, |
|
"learning_rate": 2.8e-07, |
|
"loss": 0.0599, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 4.2042042042042045, |
|
"grad_norm": 2.204902172088623, |
|
"learning_rate": 2.2444444444444445e-07, |
|
"loss": 0.0627, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 4.2256542256542255, |
|
"grad_norm": 1.7172112464904785, |
|
"learning_rate": 1.6888888888888888e-07, |
|
"loss": 0.0643, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 4.2471042471042475, |
|
"grad_norm": 1.4979685544967651, |
|
"learning_rate": 1.1333333333333336e-07, |
|
"loss": 0.0602, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 4.268554268554269, |
|
"grad_norm": 1.880300760269165, |
|
"learning_rate": 5.777777777777778e-08, |
|
"loss": 0.0591, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 4.29000429000429, |
|
"grad_norm": 1.7681268453598022, |
|
"learning_rate": 2.2222222222222225e-09, |
|
"loss": 0.0665, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.29000429000429, |
|
"eval_loss": 0.46873393654823303, |
|
"eval_runtime": 2420.0826, |
|
"eval_samples_per_second": 1.612, |
|
"eval_steps_per_second": 0.101, |
|
"eval_wer": 0.2887085868830291, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.29000429000429, |
|
"step": 5000, |
|
"total_flos": 5.435317790834688e+20, |
|
"train_loss": 0.2924629046201706, |
|
"train_runtime": 60618.8801, |
|
"train_samples_per_second": 2.639, |
|
"train_steps_per_second": 0.082 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 5000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.435317790834688e+20, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|