{ "best_metric": null, "best_model_checkpoint": null, "epoch": 15.600624024960998, "eval_steps": 1000, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.078003120124805, "grad_norm": 6.3628339767456055, "learning_rate": 5.000000000000001e-07, "loss": 1.5177, "step": 25 }, { "epoch": 0.15600624024961, "grad_norm": 5.831232070922852, "learning_rate": 9.800000000000001e-07, "loss": 1.1971, "step": 50 }, { "epoch": 0.23400936037441497, "grad_norm": 4.637149333953857, "learning_rate": 1.48e-06, "loss": 0.8901, "step": 75 }, { "epoch": 0.31201248049922, "grad_norm": 4.270649433135986, "learning_rate": 1.98e-06, "loss": 0.7938, "step": 100 }, { "epoch": 0.39001560062402496, "grad_norm": 5.095465183258057, "learning_rate": 2.4800000000000004e-06, "loss": 0.721, "step": 125 }, { "epoch": 0.46801872074882994, "grad_norm": 5.3378520011901855, "learning_rate": 2.9800000000000003e-06, "loss": 0.7105, "step": 150 }, { "epoch": 0.5460218408736349, "grad_norm": 4.5053558349609375, "learning_rate": 3.46e-06, "loss": 0.6863, "step": 175 }, { "epoch": 0.62402496099844, "grad_norm": 4.717932224273682, "learning_rate": 3.96e-06, "loss": 0.6458, "step": 200 }, { "epoch": 0.7020280811232449, "grad_norm": 4.187346458435059, "learning_rate": 4.4600000000000005e-06, "loss": 0.6414, "step": 225 }, { "epoch": 0.7800312012480499, "grad_norm": 3.789942979812622, "learning_rate": 4.960000000000001e-06, "loss": 0.6233, "step": 250 }, { "epoch": 0.858034321372855, "grad_norm": 5.197734832763672, "learning_rate": 5.460000000000001e-06, "loss": 0.601, "step": 275 }, { "epoch": 0.9360374414976599, "grad_norm": 4.216986656188965, "learning_rate": 5.9600000000000005e-06, "loss": 0.626, "step": 300 }, { "epoch": 1.0140405616224648, "grad_norm": 3.443754196166992, "learning_rate": 6.460000000000001e-06, "loss": 0.5809, "step": 325 }, { "epoch": 1.0920436817472698, "grad_norm": 4.23879861831665, "learning_rate": 6.96e-06, "loss": 0.4931, "step": 350 }, { "epoch": 1.1700468018720749, "grad_norm": 4.616464614868164, "learning_rate": 7.440000000000001e-06, "loss": 0.4833, "step": 375 }, { "epoch": 1.24804992199688, "grad_norm": 3.7008609771728516, "learning_rate": 7.94e-06, "loss": 0.4792, "step": 400 }, { "epoch": 1.3260530421216847, "grad_norm": 4.368692874908447, "learning_rate": 8.44e-06, "loss": 0.4924, "step": 425 }, { "epoch": 1.4040561622464898, "grad_norm": 3.1923470497131348, "learning_rate": 8.94e-06, "loss": 0.4968, "step": 450 }, { "epoch": 1.4820592823712948, "grad_norm": 3.799287796020508, "learning_rate": 9.440000000000001e-06, "loss": 0.4759, "step": 475 }, { "epoch": 1.5600624024960998, "grad_norm": 3.8596694469451904, "learning_rate": 9.940000000000001e-06, "loss": 0.4884, "step": 500 }, { "epoch": 1.6380655226209049, "grad_norm": 3.965977907180786, "learning_rate": 9.951111111111111e-06, "loss": 0.5001, "step": 525 }, { "epoch": 1.71606864274571, "grad_norm": 11.547295570373535, "learning_rate": 9.895555555555557e-06, "loss": 0.4972, "step": 550 }, { "epoch": 1.794071762870515, "grad_norm": 3.6599745750427246, "learning_rate": 9.84e-06, "loss": 0.4955, "step": 575 }, { "epoch": 1.8720748829953198, "grad_norm": 3.5883848667144775, "learning_rate": 9.784444444444445e-06, "loss": 0.477, "step": 600 }, { "epoch": 1.9500780031201248, "grad_norm": 3.8101539611816406, "learning_rate": 9.72888888888889e-06, "loss": 0.4649, "step": 625 }, { "epoch": 2.0280811232449296, "grad_norm": 3.238903284072876, "learning_rate": 9.673333333333334e-06, "loss": 0.4188, "step": 650 }, { "epoch": 2.1060842433697347, "grad_norm": 13.860888481140137, "learning_rate": 9.617777777777778e-06, "loss": 0.3054, "step": 675 }, { "epoch": 2.1840873634945397, "grad_norm": 3.1065433025360107, "learning_rate": 9.562222222222223e-06, "loss": 0.3189, "step": 700 }, { "epoch": 2.2620904836193447, "grad_norm": 3.5035383701324463, "learning_rate": 9.506666666666667e-06, "loss": 0.3336, "step": 725 }, { "epoch": 2.3400936037441498, "grad_norm": 3.15653395652771, "learning_rate": 9.451111111111112e-06, "loss": 0.3049, "step": 750 }, { "epoch": 2.418096723868955, "grad_norm": 3.5808379650115967, "learning_rate": 9.395555555555556e-06, "loss": 0.3039, "step": 775 }, { "epoch": 2.49609984399376, "grad_norm": 3.342785358428955, "learning_rate": 9.340000000000002e-06, "loss": 0.3245, "step": 800 }, { "epoch": 2.574102964118565, "grad_norm": 3.4457006454467773, "learning_rate": 9.284444444444444e-06, "loss": 0.3191, "step": 825 }, { "epoch": 2.6521060842433695, "grad_norm": 2.653717517852783, "learning_rate": 9.22888888888889e-06, "loss": 0.3094, "step": 850 }, { "epoch": 2.7301092043681745, "grad_norm": 3.676502227783203, "learning_rate": 9.173333333333334e-06, "loss": 0.3132, "step": 875 }, { "epoch": 2.8081123244929795, "grad_norm": 2.7381463050842285, "learning_rate": 9.117777777777778e-06, "loss": 0.325, "step": 900 }, { "epoch": 2.8861154446177846, "grad_norm": 3.100311040878296, "learning_rate": 9.062222222222224e-06, "loss": 0.3068, "step": 925 }, { "epoch": 2.9641185647425896, "grad_norm": 3.143117904663086, "learning_rate": 9.006666666666666e-06, "loss": 0.3131, "step": 950 }, { "epoch": 3.0421216848673946, "grad_norm": 2.2605462074279785, "learning_rate": 8.951111111111112e-06, "loss": 0.2532, "step": 975 }, { "epoch": 3.1201248049921997, "grad_norm": 2.582007884979248, "learning_rate": 8.895555555555556e-06, "loss": 0.18, "step": 1000 }, { "epoch": 3.1201248049921997, "eval_loss": 0.5584835410118103, "eval_runtime": 521.3754, "eval_samples_per_second": 2.457, "eval_steps_per_second": 0.155, "eval_wer": 0.37714912815510304, "step": 1000 }, { "epoch": 3.1981279251170047, "grad_norm": 2.988361120223999, "learning_rate": 8.84e-06, "loss": 0.1874, "step": 1025 }, { "epoch": 3.2761310452418098, "grad_norm": 4.793211936950684, "learning_rate": 8.784444444444446e-06, "loss": 0.1892, "step": 1050 }, { "epoch": 3.354134165366615, "grad_norm": 2.9947147369384766, "learning_rate": 8.72888888888889e-06, "loss": 0.1888, "step": 1075 }, { "epoch": 3.43213728549142, "grad_norm": 3.2537567615509033, "learning_rate": 8.673333333333334e-06, "loss": 0.1949, "step": 1100 }, { "epoch": 3.510140405616225, "grad_norm": 2.6582891941070557, "learning_rate": 8.617777777777778e-06, "loss": 0.1933, "step": 1125 }, { "epoch": 3.58814352574103, "grad_norm": 2.752394914627075, "learning_rate": 8.562222222222224e-06, "loss": 0.1967, "step": 1150 }, { "epoch": 3.6661466458658345, "grad_norm": 2.853532552719116, "learning_rate": 8.506666666666668e-06, "loss": 0.1978, "step": 1175 }, { "epoch": 3.7441497659906395, "grad_norm": 2.4331047534942627, "learning_rate": 8.451111111111112e-06, "loss": 0.2051, "step": 1200 }, { "epoch": 3.8221528861154446, "grad_norm": 2.7613861560821533, "learning_rate": 8.395555555555557e-06, "loss": 0.1948, "step": 1225 }, { "epoch": 3.9001560062402496, "grad_norm": 2.684624433517456, "learning_rate": 8.34e-06, "loss": 0.2048, "step": 1250 }, { "epoch": 3.9781591263650546, "grad_norm": 4.769908428192139, "learning_rate": 8.284444444444446e-06, "loss": 0.2045, "step": 1275 }, { "epoch": 4.056162246489859, "grad_norm": 2.2372353076934814, "learning_rate": 8.22888888888889e-06, "loss": 0.1319, "step": 1300 }, { "epoch": 4.134165366614664, "grad_norm": 7.478884220123291, "learning_rate": 8.173333333333334e-06, "loss": 0.1106, "step": 1325 }, { "epoch": 4.212168486739469, "grad_norm": 3.3369452953338623, "learning_rate": 8.11777777777778e-06, "loss": 0.1115, "step": 1350 }, { "epoch": 4.290171606864274, "grad_norm": 2.1033334732055664, "learning_rate": 8.062222222222222e-06, "loss": 0.1148, "step": 1375 }, { "epoch": 4.368174726989079, "grad_norm": 1.9185432195663452, "learning_rate": 8.006666666666667e-06, "loss": 0.1082, "step": 1400 }, { "epoch": 4.446177847113884, "grad_norm": 2.428751230239868, "learning_rate": 7.951111111111111e-06, "loss": 0.1129, "step": 1425 }, { "epoch": 4.5241809672386895, "grad_norm": 2.757439374923706, "learning_rate": 7.895555555555557e-06, "loss": 0.1184, "step": 1450 }, { "epoch": 4.6021840873634945, "grad_norm": 2.134047508239746, "learning_rate": 7.840000000000001e-06, "loss": 0.1109, "step": 1475 }, { "epoch": 4.6801872074882995, "grad_norm": 3.3976402282714844, "learning_rate": 7.784444444444445e-06, "loss": 0.1209, "step": 1500 }, { "epoch": 4.758190327613105, "grad_norm": 2.2624588012695312, "learning_rate": 7.72888888888889e-06, "loss": 0.116, "step": 1525 }, { "epoch": 4.83619344773791, "grad_norm": 2.8306870460510254, "learning_rate": 7.673333333333333e-06, "loss": 0.1174, "step": 1550 }, { "epoch": 4.914196567862715, "grad_norm": 2.3666274547576904, "learning_rate": 7.617777777777778e-06, "loss": 0.1127, "step": 1575 }, { "epoch": 4.99219968798752, "grad_norm": 2.4004135131835938, "learning_rate": 7.562222222222223e-06, "loss": 0.1143, "step": 1600 }, { "epoch": 5.070202808112325, "grad_norm": 1.8675142526626587, "learning_rate": 7.506666666666668e-06, "loss": 0.0694, "step": 1625 }, { "epoch": 5.14820592823713, "grad_norm": 1.7827036380767822, "learning_rate": 7.451111111111111e-06, "loss": 0.0613, "step": 1650 }, { "epoch": 5.226209048361935, "grad_norm": 1.6281778812408447, "learning_rate": 7.395555555555556e-06, "loss": 0.062, "step": 1675 }, { "epoch": 5.30421216848674, "grad_norm": 3.8592145442962646, "learning_rate": 7.340000000000001e-06, "loss": 0.0672, "step": 1700 }, { "epoch": 5.382215288611545, "grad_norm": 2.766690731048584, "learning_rate": 7.284444444444445e-06, "loss": 0.0594, "step": 1725 }, { "epoch": 5.460218408736349, "grad_norm": 3.115182399749756, "learning_rate": 7.22888888888889e-06, "loss": 0.0625, "step": 1750 }, { "epoch": 5.538221528861154, "grad_norm": 1.6196472644805908, "learning_rate": 7.173333333333335e-06, "loss": 0.06, "step": 1775 }, { "epoch": 5.616224648985959, "grad_norm": 2.140605926513672, "learning_rate": 7.117777777777778e-06, "loss": 0.0633, "step": 1800 }, { "epoch": 5.694227769110764, "grad_norm": 2.190873622894287, "learning_rate": 7.062222222222223e-06, "loss": 0.0678, "step": 1825 }, { "epoch": 5.772230889235569, "grad_norm": 2.369245767593384, "learning_rate": 7.006666666666667e-06, "loss": 0.0657, "step": 1850 }, { "epoch": 5.850234009360374, "grad_norm": 1.8441182374954224, "learning_rate": 6.951111111111112e-06, "loss": 0.0652, "step": 1875 }, { "epoch": 5.928237129485179, "grad_norm": 1.8645758628845215, "learning_rate": 6.8955555555555565e-06, "loss": 0.0643, "step": 1900 }, { "epoch": 6.006240249609984, "grad_norm": 2.1204230785369873, "learning_rate": 6.8400000000000014e-06, "loss": 0.0614, "step": 1925 }, { "epoch": 6.084243369734789, "grad_norm": 1.5469348430633545, "learning_rate": 6.784444444444445e-06, "loss": 0.0364, "step": 1950 }, { "epoch": 6.162246489859594, "grad_norm": 2.485136032104492, "learning_rate": 6.7288888888888895e-06, "loss": 0.0356, "step": 1975 }, { "epoch": 6.240249609984399, "grad_norm": 1.3276453018188477, "learning_rate": 6.6733333333333335e-06, "loss": 0.0328, "step": 2000 }, { "epoch": 6.240249609984399, "eval_loss": 0.6914765238761902, "eval_runtime": 519.148, "eval_samples_per_second": 2.468, "eval_steps_per_second": 0.156, "eval_wer": 0.3699548835507865, "step": 2000 }, { "epoch": 6.318252730109204, "grad_norm": 2.1222565174102783, "learning_rate": 6.617777777777778e-06, "loss": 0.0363, "step": 2025 }, { "epoch": 6.396255850234009, "grad_norm": 1.7250409126281738, "learning_rate": 6.562222222222223e-06, "loss": 0.036, "step": 2050 }, { "epoch": 6.4742589703588145, "grad_norm": 1.5123999118804932, "learning_rate": 6.5066666666666665e-06, "loss": 0.0382, "step": 2075 }, { "epoch": 6.5522620904836195, "grad_norm": 1.8385040760040283, "learning_rate": 6.451111111111111e-06, "loss": 0.0362, "step": 2100 }, { "epoch": 6.6302652106084246, "grad_norm": 1.9721713066101074, "learning_rate": 6.395555555555556e-06, "loss": 0.0374, "step": 2125 }, { "epoch": 6.70826833073323, "grad_norm": 2.1653904914855957, "learning_rate": 6.34e-06, "loss": 0.0372, "step": 2150 }, { "epoch": 6.786271450858035, "grad_norm": 1.9096741676330566, "learning_rate": 6.284444444444445e-06, "loss": 0.0362, "step": 2175 }, { "epoch": 6.86427457098284, "grad_norm": 1.5271517038345337, "learning_rate": 6.22888888888889e-06, "loss": 0.0375, "step": 2200 }, { "epoch": 6.942277691107645, "grad_norm": 1.6990188360214233, "learning_rate": 6.173333333333333e-06, "loss": 0.0377, "step": 2225 }, { "epoch": 7.02028081123245, "grad_norm": 1.4633820056915283, "learning_rate": 6.117777777777778e-06, "loss": 0.0362, "step": 2250 }, { "epoch": 7.098283931357254, "grad_norm": 0.841683566570282, "learning_rate": 6.062222222222223e-06, "loss": 0.0197, "step": 2275 }, { "epoch": 7.176287051482059, "grad_norm": 1.274117112159729, "learning_rate": 6.006666666666667e-06, "loss": 0.0203, "step": 2300 }, { "epoch": 7.254290171606864, "grad_norm": 2.547511100769043, "learning_rate": 5.951111111111112e-06, "loss": 0.0232, "step": 2325 }, { "epoch": 7.332293291731669, "grad_norm": 1.2308272123336792, "learning_rate": 5.895555555555557e-06, "loss": 0.0248, "step": 2350 }, { "epoch": 7.410296411856474, "grad_norm": 1.271571159362793, "learning_rate": 5.84e-06, "loss": 0.0204, "step": 2375 }, { "epoch": 7.488299531981279, "grad_norm": 2.4256279468536377, "learning_rate": 5.784444444444445e-06, "loss": 0.0236, "step": 2400 }, { "epoch": 7.566302652106084, "grad_norm": 1.8029208183288574, "learning_rate": 5.72888888888889e-06, "loss": 0.0225, "step": 2425 }, { "epoch": 7.644305772230889, "grad_norm": 1.7934774160385132, "learning_rate": 5.673333333333334e-06, "loss": 0.0222, "step": 2450 }, { "epoch": 7.722308892355694, "grad_norm": 0.9680442214012146, "learning_rate": 5.617777777777779e-06, "loss": 0.0219, "step": 2475 }, { "epoch": 7.800312012480499, "grad_norm": 1.29971182346344, "learning_rate": 5.562222222222222e-06, "loss": 0.0226, "step": 2500 }, { "epoch": 7.878315132605304, "grad_norm": 1.8278181552886963, "learning_rate": 5.506666666666667e-06, "loss": 0.0236, "step": 2525 }, { "epoch": 7.956318252730109, "grad_norm": 1.8502779006958008, "learning_rate": 5.451111111111112e-06, "loss": 0.0229, "step": 2550 }, { "epoch": 8.034321372854913, "grad_norm": 2.3189539909362793, "learning_rate": 5.3955555555555565e-06, "loss": 0.0193, "step": 2575 }, { "epoch": 8.112324492979718, "grad_norm": 0.7720842361450195, "learning_rate": 5.3400000000000005e-06, "loss": 0.0121, "step": 2600 }, { "epoch": 8.190327613104524, "grad_norm": 3.5574843883514404, "learning_rate": 5.2844444444444454e-06, "loss": 0.0155, "step": 2625 }, { "epoch": 8.268330733229329, "grad_norm": 1.5664520263671875, "learning_rate": 5.228888888888889e-06, "loss": 0.0147, "step": 2650 }, { "epoch": 8.346333853354134, "grad_norm": 1.2214146852493286, "learning_rate": 5.1733333333333335e-06, "loss": 0.0149, "step": 2675 }, { "epoch": 8.424336973478939, "grad_norm": 1.0644782781600952, "learning_rate": 5.117777777777778e-06, "loss": 0.0131, "step": 2700 }, { "epoch": 8.502340093603744, "grad_norm": 0.8796694278717041, "learning_rate": 5.062222222222222e-06, "loss": 0.014, "step": 2725 }, { "epoch": 8.580343213728549, "grad_norm": 0.8544178009033203, "learning_rate": 5.006666666666667e-06, "loss": 0.0146, "step": 2750 }, { "epoch": 8.658346333853354, "grad_norm": 0.7999880313873291, "learning_rate": 4.951111111111111e-06, "loss": 0.0147, "step": 2775 }, { "epoch": 8.736349453978159, "grad_norm": 1.7287328243255615, "learning_rate": 4.895555555555556e-06, "loss": 0.0148, "step": 2800 }, { "epoch": 8.814352574102964, "grad_norm": 1.2320371866226196, "learning_rate": 4.84e-06, "loss": 0.0117, "step": 2825 }, { "epoch": 8.892355694227769, "grad_norm": 0.6565980315208435, "learning_rate": 4.784444444444445e-06, "loss": 0.0171, "step": 2850 }, { "epoch": 8.970358814352574, "grad_norm": 0.6645917296409607, "learning_rate": 4.728888888888889e-06, "loss": 0.0146, "step": 2875 }, { "epoch": 9.048361934477379, "grad_norm": 1.5899229049682617, "learning_rate": 4.673333333333333e-06, "loss": 0.0095, "step": 2900 }, { "epoch": 9.126365054602184, "grad_norm": 1.3482928276062012, "learning_rate": 4.617777777777778e-06, "loss": 0.0102, "step": 2925 }, { "epoch": 9.204368174726989, "grad_norm": 1.0298078060150146, "learning_rate": 4.562222222222222e-06, "loss": 0.01, "step": 2950 }, { "epoch": 9.282371294851794, "grad_norm": 0.7524386048316956, "learning_rate": 4.506666666666667e-06, "loss": 0.0087, "step": 2975 }, { "epoch": 9.360374414976599, "grad_norm": 1.3448346853256226, "learning_rate": 4.451111111111112e-06, "loss": 0.0086, "step": 3000 }, { "epoch": 9.360374414976599, "eval_loss": 0.8110594749450684, "eval_runtime": 514.7694, "eval_samples_per_second": 2.488, "eval_steps_per_second": 0.157, "eval_wer": 0.3620899890257286, "step": 3000 }, { "epoch": 9.438377535101404, "grad_norm": 1.4639930725097656, "learning_rate": 4.395555555555556e-06, "loss": 0.0089, "step": 3025 }, { "epoch": 9.51638065522621, "grad_norm": 0.7041357755661011, "learning_rate": 4.34e-06, "loss": 0.01, "step": 3050 }, { "epoch": 9.594383775351014, "grad_norm": 0.8498127460479736, "learning_rate": 4.284444444444445e-06, "loss": 0.0085, "step": 3075 }, { "epoch": 9.67238689547582, "grad_norm": 0.8553516864776611, "learning_rate": 4.228888888888889e-06, "loss": 0.0101, "step": 3100 }, { "epoch": 9.750390015600624, "grad_norm": 2.619405746459961, "learning_rate": 4.173333333333334e-06, "loss": 0.0099, "step": 3125 }, { "epoch": 9.82839313572543, "grad_norm": 1.0941424369812012, "learning_rate": 4.117777777777779e-06, "loss": 0.0089, "step": 3150 }, { "epoch": 9.906396255850234, "grad_norm": 1.0034147500991821, "learning_rate": 4.062222222222223e-06, "loss": 0.0087, "step": 3175 }, { "epoch": 9.98439937597504, "grad_norm": 0.6879417896270752, "learning_rate": 4.006666666666667e-06, "loss": 0.0086, "step": 3200 }, { "epoch": 10.062402496099844, "grad_norm": 0.5009737014770508, "learning_rate": 3.951111111111112e-06, "loss": 0.0051, "step": 3225 }, { "epoch": 10.14040561622465, "grad_norm": 0.3143404424190521, "learning_rate": 3.895555555555556e-06, "loss": 0.0048, "step": 3250 }, { "epoch": 10.218408736349454, "grad_norm": 0.44424840807914734, "learning_rate": 3.8400000000000005e-06, "loss": 0.0046, "step": 3275 }, { "epoch": 10.29641185647426, "grad_norm": 0.18599098920822144, "learning_rate": 3.784444444444445e-06, "loss": 0.0039, "step": 3300 }, { "epoch": 10.374414976599065, "grad_norm": 0.2974778115749359, "learning_rate": 3.728888888888889e-06, "loss": 0.0046, "step": 3325 }, { "epoch": 10.45241809672387, "grad_norm": 0.990300178527832, "learning_rate": 3.673333333333334e-06, "loss": 0.0041, "step": 3350 }, { "epoch": 10.530421216848675, "grad_norm": 1.2518631219863892, "learning_rate": 3.617777777777778e-06, "loss": 0.0048, "step": 3375 }, { "epoch": 10.60842433697348, "grad_norm": 0.9894776940345764, "learning_rate": 3.5622222222222224e-06, "loss": 0.0053, "step": 3400 }, { "epoch": 10.686427457098285, "grad_norm": 0.192794531583786, "learning_rate": 3.5066666666666673e-06, "loss": 0.0059, "step": 3425 }, { "epoch": 10.76443057722309, "grad_norm": 0.30634793639183044, "learning_rate": 3.4511111111111113e-06, "loss": 0.0055, "step": 3450 }, { "epoch": 10.842433697347893, "grad_norm": 0.2689841687679291, "learning_rate": 3.3955555555555558e-06, "loss": 0.0051, "step": 3475 }, { "epoch": 10.920436817472698, "grad_norm": 2.5046868324279785, "learning_rate": 3.3400000000000006e-06, "loss": 0.0038, "step": 3500 }, { "epoch": 10.998439937597503, "grad_norm": 0.7331063151359558, "learning_rate": 3.2844444444444447e-06, "loss": 0.0046, "step": 3525 }, { "epoch": 11.076443057722308, "grad_norm": 0.3739769458770752, "learning_rate": 3.228888888888889e-06, "loss": 0.0033, "step": 3550 }, { "epoch": 11.154446177847113, "grad_norm": 0.4729616343975067, "learning_rate": 3.173333333333334e-06, "loss": 0.0036, "step": 3575 }, { "epoch": 11.232449297971918, "grad_norm": 0.15952661633491516, "learning_rate": 3.117777777777778e-06, "loss": 0.0029, "step": 3600 }, { "epoch": 11.310452418096723, "grad_norm": 0.20027977228164673, "learning_rate": 3.0622222222222225e-06, "loss": 0.0036, "step": 3625 }, { "epoch": 11.388455538221528, "grad_norm": 0.19324319064617157, "learning_rate": 3.0066666666666674e-06, "loss": 0.0033, "step": 3650 }, { "epoch": 11.466458658346333, "grad_norm": 0.20121298730373383, "learning_rate": 2.9511111111111114e-06, "loss": 0.0022, "step": 3675 }, { "epoch": 11.544461778471138, "grad_norm": 0.7968800663948059, "learning_rate": 2.895555555555556e-06, "loss": 0.0025, "step": 3700 }, { "epoch": 11.622464898595943, "grad_norm": 0.1721133291721344, "learning_rate": 2.84e-06, "loss": 0.0026, "step": 3725 }, { "epoch": 11.700468018720748, "grad_norm": 0.37284353375434875, "learning_rate": 2.784444444444445e-06, "loss": 0.0028, "step": 3750 }, { "epoch": 11.778471138845553, "grad_norm": 0.1511831283569336, "learning_rate": 2.7288888888888893e-06, "loss": 0.0023, "step": 3775 }, { "epoch": 11.856474258970358, "grad_norm": 0.10786991566419601, "learning_rate": 2.6733333333333333e-06, "loss": 0.0029, "step": 3800 }, { "epoch": 11.934477379095163, "grad_norm": 1.2048157453536987, "learning_rate": 2.617777777777778e-06, "loss": 0.0032, "step": 3825 }, { "epoch": 12.012480499219969, "grad_norm": 0.1635945737361908, "learning_rate": 2.5622222222222226e-06, "loss": 0.0026, "step": 3850 }, { "epoch": 12.090483619344774, "grad_norm": 0.06904599815607071, "learning_rate": 2.5066666666666667e-06, "loss": 0.0026, "step": 3875 }, { "epoch": 12.168486739469579, "grad_norm": 0.06420668214559555, "learning_rate": 2.451111111111111e-06, "loss": 0.002, "step": 3900 }, { "epoch": 12.246489859594384, "grad_norm": 0.26706597208976746, "learning_rate": 2.3955555555555556e-06, "loss": 0.0025, "step": 3925 }, { "epoch": 12.324492979719189, "grad_norm": 0.05851108580827713, "learning_rate": 2.3400000000000005e-06, "loss": 0.0018, "step": 3950 }, { "epoch": 12.402496099843994, "grad_norm": 0.06494307518005371, "learning_rate": 2.2844444444444445e-06, "loss": 0.0015, "step": 3975 }, { "epoch": 12.480499219968799, "grad_norm": 0.07140354067087173, "learning_rate": 2.228888888888889e-06, "loss": 0.0017, "step": 4000 }, { "epoch": 12.480499219968799, "eval_loss": 0.8715334534645081, "eval_runtime": 513.4704, "eval_samples_per_second": 2.495, "eval_steps_per_second": 0.158, "eval_wer": 0.360260943787343, "step": 4000 }, { "epoch": 12.558502340093604, "grad_norm": 0.04837729409337044, "learning_rate": 2.1733333333333334e-06, "loss": 0.0014, "step": 4025 }, { "epoch": 12.636505460218409, "grad_norm": 0.050302207469940186, "learning_rate": 2.117777777777778e-06, "loss": 0.0015, "step": 4050 }, { "epoch": 12.714508580343214, "grad_norm": 1.5353797674179077, "learning_rate": 2.0622222222222223e-06, "loss": 0.0019, "step": 4075 }, { "epoch": 12.792511700468019, "grad_norm": 0.0588994063436985, "learning_rate": 2.006666666666667e-06, "loss": 0.0017, "step": 4100 }, { "epoch": 12.870514820592824, "grad_norm": 0.05936579406261444, "learning_rate": 1.9511111111111113e-06, "loss": 0.002, "step": 4125 }, { "epoch": 12.948517940717629, "grad_norm": 0.1404724270105362, "learning_rate": 1.8955555555555557e-06, "loss": 0.0023, "step": 4150 }, { "epoch": 13.026521060842434, "grad_norm": 0.07803475856781006, "learning_rate": 1.8400000000000002e-06, "loss": 0.0013, "step": 4175 }, { "epoch": 13.104524180967239, "grad_norm": 0.045549679547548294, "learning_rate": 1.7844444444444444e-06, "loss": 0.0013, "step": 4200 }, { "epoch": 13.182527301092044, "grad_norm": 1.5323783159255981, "learning_rate": 1.728888888888889e-06, "loss": 0.0018, "step": 4225 }, { "epoch": 13.260530421216849, "grad_norm": 0.0453680157661438, "learning_rate": 1.6733333333333335e-06, "loss": 0.0011, "step": 4250 }, { "epoch": 13.338533541341654, "grad_norm": 0.16936659812927246, "learning_rate": 1.6177777777777778e-06, "loss": 0.0012, "step": 4275 }, { "epoch": 13.41653666146646, "grad_norm": 0.03871113434433937, "learning_rate": 1.5622222222222225e-06, "loss": 0.0013, "step": 4300 }, { "epoch": 13.494539781591264, "grad_norm": 0.10738777369260788, "learning_rate": 1.506666666666667e-06, "loss": 0.0011, "step": 4325 }, { "epoch": 13.57254290171607, "grad_norm": 0.047905419021844864, "learning_rate": 1.4511111111111112e-06, "loss": 0.001, "step": 4350 }, { "epoch": 13.650546021840874, "grad_norm": 0.04242739453911781, "learning_rate": 1.3955555555555556e-06, "loss": 0.0013, "step": 4375 }, { "epoch": 13.72854914196568, "grad_norm": 0.04559892416000366, "learning_rate": 1.34e-06, "loss": 0.0012, "step": 4400 }, { "epoch": 13.806552262090484, "grad_norm": 0.030529584735631943, "learning_rate": 1.2844444444444445e-06, "loss": 0.0011, "step": 4425 }, { "epoch": 13.88455538221529, "grad_norm": 0.03567594662308693, "learning_rate": 1.228888888888889e-06, "loss": 0.001, "step": 4450 }, { "epoch": 13.962558502340094, "grad_norm": 0.04149959608912468, "learning_rate": 1.1733333333333335e-06, "loss": 0.0011, "step": 4475 }, { "epoch": 14.0405616224649, "grad_norm": 0.0304497629404068, "learning_rate": 1.117777777777778e-06, "loss": 0.0009, "step": 4500 }, { "epoch": 14.118564742589703, "grad_norm": 0.028642022982239723, "learning_rate": 1.0622222222222222e-06, "loss": 0.0009, "step": 4525 }, { "epoch": 14.196567862714508, "grad_norm": 0.02880062349140644, "learning_rate": 1.0066666666666668e-06, "loss": 0.0009, "step": 4550 }, { "epoch": 14.274570982839313, "grad_norm": 0.03172925487160683, "learning_rate": 9.511111111111111e-07, "loss": 0.0014, "step": 4575 }, { "epoch": 14.352574102964118, "grad_norm": 0.02901492826640606, "learning_rate": 8.955555555555557e-07, "loss": 0.0009, "step": 4600 }, { "epoch": 14.430577223088923, "grad_norm": 0.027313807979226112, "learning_rate": 8.400000000000001e-07, "loss": 0.0009, "step": 4625 }, { "epoch": 14.508580343213728, "grad_norm": 0.03275463357567787, "learning_rate": 7.844444444444445e-07, "loss": 0.0011, "step": 4650 }, { "epoch": 14.586583463338533, "grad_norm": 0.02867468073964119, "learning_rate": 7.28888888888889e-07, "loss": 0.0009, "step": 4675 }, { "epoch": 14.664586583463338, "grad_norm": 0.02795287035405636, "learning_rate": 6.733333333333334e-07, "loss": 0.0009, "step": 4700 }, { "epoch": 14.742589703588143, "grad_norm": 0.033373042941093445, "learning_rate": 6.177777777777778e-07, "loss": 0.0009, "step": 4725 }, { "epoch": 14.820592823712948, "grad_norm": 0.10227759182453156, "learning_rate": 5.622222222222223e-07, "loss": 0.001, "step": 4750 }, { "epoch": 14.898595943837753, "grad_norm": 0.03362324461340904, "learning_rate": 5.066666666666667e-07, "loss": 0.0009, "step": 4775 }, { "epoch": 14.976599063962558, "grad_norm": 0.02761083096265793, "learning_rate": 4.511111111111111e-07, "loss": 0.0009, "step": 4800 }, { "epoch": 15.054602184087363, "grad_norm": 0.030326733365654945, "learning_rate": 3.9555555555555557e-07, "loss": 0.0009, "step": 4825 }, { "epoch": 15.132605304212168, "grad_norm": 0.02800353430211544, "learning_rate": 3.4000000000000003e-07, "loss": 0.0008, "step": 4850 }, { "epoch": 15.210608424336973, "grad_norm": 0.033761948347091675, "learning_rate": 2.844444444444445e-07, "loss": 0.0008, "step": 4875 }, { "epoch": 15.288611544461778, "grad_norm": 0.030224645510315895, "learning_rate": 2.2888888888888892e-07, "loss": 0.0008, "step": 4900 }, { "epoch": 15.366614664586583, "grad_norm": 0.023724382743239403, "learning_rate": 1.7333333333333335e-07, "loss": 0.0008, "step": 4925 }, { "epoch": 15.444617784711388, "grad_norm": 0.03140026330947876, "learning_rate": 1.1777777777777778e-07, "loss": 0.0009, "step": 4950 }, { "epoch": 15.522620904836193, "grad_norm": 0.03177293762564659, "learning_rate": 6.222222222222223e-08, "loss": 0.0011, "step": 4975 }, { "epoch": 15.600624024960998, "grad_norm": 0.023135744035243988, "learning_rate": 6.666666666666667e-09, "loss": 0.0008, "step": 5000 }, { "epoch": 15.600624024960998, "eval_loss": 0.9218041896820068, "eval_runtime": 514.1223, "eval_samples_per_second": 2.492, "eval_steps_per_second": 0.158, "eval_wer": 0.35574929886599194, "step": 5000 }, { "epoch": 15.600624024960998, "step": 5000, "total_flos": 3.39293344978944e+20, "train_loss": 0.13203169204816223, "train_runtime": 41579.1778, "train_samples_per_second": 3.848, "train_steps_per_second": 0.12 } ], "logging_steps": 25, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 16, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.39293344978944e+20, "train_batch_size": 16, "trial_name": null, "trial_params": null }