whisper-medium-mix-it / trainer_state.json
deepdml's picture
End of training
25fb756 verified
{
"best_metric": 6.840122206312234,
"best_model_checkpoint": "./whisper-medium-mix-it/checkpoint-5000",
"epoch": 1.0,
"eval_steps": 1000,
"global_step": 5000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.005,
"grad_norm": 7.006679058074951,
"learning_rate": 4.6000000000000004e-07,
"loss": 0.7121,
"step": 25
},
{
"epoch": 0.01,
"grad_norm": 4.37919282913208,
"learning_rate": 9.600000000000001e-07,
"loss": 0.5421,
"step": 50
},
{
"epoch": 0.015,
"grad_norm": 3.432912826538086,
"learning_rate": 1.46e-06,
"loss": 0.3692,
"step": 75
},
{
"epoch": 0.02,
"grad_norm": 2.942035436630249,
"learning_rate": 1.9600000000000003e-06,
"loss": 0.2601,
"step": 100
},
{
"epoch": 0.025,
"grad_norm": 2.9149224758148193,
"learning_rate": 2.46e-06,
"loss": 0.2557,
"step": 125
},
{
"epoch": 0.03,
"grad_norm": 2.536968231201172,
"learning_rate": 2.96e-06,
"loss": 0.216,
"step": 150
},
{
"epoch": 0.035,
"grad_norm": 4.307021617889404,
"learning_rate": 3.46e-06,
"loss": 0.2004,
"step": 175
},
{
"epoch": 0.04,
"grad_norm": 3.456475019454956,
"learning_rate": 3.96e-06,
"loss": 0.2118,
"step": 200
},
{
"epoch": 0.045,
"grad_norm": 3.5058531761169434,
"learning_rate": 4.4600000000000005e-06,
"loss": 0.2071,
"step": 225
},
{
"epoch": 0.05,
"grad_norm": 3.056401014328003,
"learning_rate": 4.960000000000001e-06,
"loss": 0.2103,
"step": 250
},
{
"epoch": 0.055,
"grad_norm": 3.7942306995391846,
"learning_rate": 5.460000000000001e-06,
"loss": 0.2138,
"step": 275
},
{
"epoch": 0.06,
"grad_norm": 3.421769857406616,
"learning_rate": 5.9600000000000005e-06,
"loss": 0.2114,
"step": 300
},
{
"epoch": 0.065,
"grad_norm": 3.5847620964050293,
"learning_rate": 6.460000000000001e-06,
"loss": 0.1987,
"step": 325
},
{
"epoch": 0.07,
"grad_norm": 3.585036039352417,
"learning_rate": 6.96e-06,
"loss": 0.1886,
"step": 350
},
{
"epoch": 0.075,
"grad_norm": 3.3235092163085938,
"learning_rate": 7.4600000000000006e-06,
"loss": 0.1883,
"step": 375
},
{
"epoch": 0.08,
"grad_norm": 4.223800182342529,
"learning_rate": 7.960000000000002e-06,
"loss": 0.2115,
"step": 400
},
{
"epoch": 0.085,
"grad_norm": 3.307344675064087,
"learning_rate": 8.46e-06,
"loss": 0.1942,
"step": 425
},
{
"epoch": 0.09,
"grad_norm": 3.1430752277374268,
"learning_rate": 8.96e-06,
"loss": 0.1796,
"step": 450
},
{
"epoch": 0.095,
"grad_norm": 3.7946882247924805,
"learning_rate": 9.460000000000001e-06,
"loss": 0.2053,
"step": 475
},
{
"epoch": 0.1,
"grad_norm": 3.383258581161499,
"learning_rate": 9.960000000000001e-06,
"loss": 0.1918,
"step": 500
},
{
"epoch": 0.105,
"grad_norm": 3.1705129146575928,
"learning_rate": 9.94888888888889e-06,
"loss": 0.2061,
"step": 525
},
{
"epoch": 0.11,
"grad_norm": 3.299571990966797,
"learning_rate": 9.893333333333334e-06,
"loss": 0.1891,
"step": 550
},
{
"epoch": 0.115,
"grad_norm": 3.413519859313965,
"learning_rate": 9.837777777777778e-06,
"loss": 0.2221,
"step": 575
},
{
"epoch": 0.12,
"grad_norm": 3.8573830127716064,
"learning_rate": 9.782222222222222e-06,
"loss": 0.2092,
"step": 600
},
{
"epoch": 0.125,
"grad_norm": 2.475337266921997,
"learning_rate": 9.726666666666668e-06,
"loss": 0.2199,
"step": 625
},
{
"epoch": 0.13,
"grad_norm": 2.221552610397339,
"learning_rate": 9.671111111111112e-06,
"loss": 0.1787,
"step": 650
},
{
"epoch": 0.135,
"grad_norm": 3.197284460067749,
"learning_rate": 9.615555555555558e-06,
"loss": 0.2044,
"step": 675
},
{
"epoch": 0.14,
"grad_norm": 3.177967071533203,
"learning_rate": 9.56e-06,
"loss": 0.196,
"step": 700
},
{
"epoch": 0.145,
"grad_norm": 3.4896106719970703,
"learning_rate": 9.504444444444446e-06,
"loss": 0.1887,
"step": 725
},
{
"epoch": 0.15,
"grad_norm": 3.654257297515869,
"learning_rate": 9.44888888888889e-06,
"loss": 0.1979,
"step": 750
},
{
"epoch": 0.155,
"grad_norm": 3.537205934524536,
"learning_rate": 9.393333333333334e-06,
"loss": 0.1855,
"step": 775
},
{
"epoch": 0.16,
"grad_norm": 2.230680227279663,
"learning_rate": 9.33777777777778e-06,
"loss": 0.1668,
"step": 800
},
{
"epoch": 0.165,
"grad_norm": 2.3759679794311523,
"learning_rate": 9.282222222222222e-06,
"loss": 0.1643,
"step": 825
},
{
"epoch": 0.17,
"grad_norm": 2.8923466205596924,
"learning_rate": 9.226666666666668e-06,
"loss": 0.1781,
"step": 850
},
{
"epoch": 0.175,
"grad_norm": 2.956648349761963,
"learning_rate": 9.171111111111112e-06,
"loss": 0.1708,
"step": 875
},
{
"epoch": 0.18,
"grad_norm": 2.7545344829559326,
"learning_rate": 9.115555555555556e-06,
"loss": 0.161,
"step": 900
},
{
"epoch": 0.185,
"grad_norm": 2.0833823680877686,
"learning_rate": 9.060000000000001e-06,
"loss": 0.1576,
"step": 925
},
{
"epoch": 0.19,
"grad_norm": 2.689265727996826,
"learning_rate": 9.004444444444445e-06,
"loss": 0.1702,
"step": 950
},
{
"epoch": 0.195,
"grad_norm": 2.9991567134857178,
"learning_rate": 8.94888888888889e-06,
"loss": 0.1571,
"step": 975
},
{
"epoch": 0.2,
"grad_norm": 2.6324071884155273,
"learning_rate": 8.893333333333333e-06,
"loss": 0.1502,
"step": 1000
},
{
"epoch": 0.2,
"eval_loss": 0.17080245912075043,
"eval_runtime": 1706.7457,
"eval_samples_per_second": 8.879,
"eval_steps_per_second": 1.11,
"eval_wer": 9.092228813503402,
"step": 1000
},
{
"epoch": 0.205,
"grad_norm": 3.2417123317718506,
"learning_rate": 8.83777777777778e-06,
"loss": 0.1723,
"step": 1025
},
{
"epoch": 0.21,
"grad_norm": 4.017546653747559,
"learning_rate": 8.782222222222223e-06,
"loss": 0.1652,
"step": 1050
},
{
"epoch": 0.215,
"grad_norm": 2.1816272735595703,
"learning_rate": 8.726666666666667e-06,
"loss": 0.1738,
"step": 1075
},
{
"epoch": 0.22,
"grad_norm": 2.869568347930908,
"learning_rate": 8.671111111111113e-06,
"loss": 0.1975,
"step": 1100
},
{
"epoch": 0.225,
"grad_norm": 3.047701835632324,
"learning_rate": 8.615555555555555e-06,
"loss": 0.2191,
"step": 1125
},
{
"epoch": 0.23,
"grad_norm": 3.380408763885498,
"learning_rate": 8.560000000000001e-06,
"loss": 0.2155,
"step": 1150
},
{
"epoch": 0.235,
"grad_norm": 2.9988808631896973,
"learning_rate": 8.504444444444445e-06,
"loss": 0.2288,
"step": 1175
},
{
"epoch": 0.24,
"grad_norm": 2.327162027359009,
"learning_rate": 8.448888888888889e-06,
"loss": 0.2262,
"step": 1200
},
{
"epoch": 0.245,
"grad_norm": 2.988764762878418,
"learning_rate": 8.393333333333335e-06,
"loss": 0.191,
"step": 1225
},
{
"epoch": 0.25,
"grad_norm": 3.3341264724731445,
"learning_rate": 8.337777777777777e-06,
"loss": 0.1974,
"step": 1250
},
{
"epoch": 0.255,
"grad_norm": 3.1267290115356445,
"learning_rate": 8.282222222222223e-06,
"loss": 0.1951,
"step": 1275
},
{
"epoch": 0.26,
"grad_norm": 2.5510425567626953,
"learning_rate": 8.226666666666667e-06,
"loss": 0.1787,
"step": 1300
},
{
"epoch": 0.265,
"grad_norm": 2.5202274322509766,
"learning_rate": 8.171111111111113e-06,
"loss": 0.1549,
"step": 1325
},
{
"epoch": 0.27,
"grad_norm": 2.8440310955047607,
"learning_rate": 8.115555555555557e-06,
"loss": 0.1858,
"step": 1350
},
{
"epoch": 0.275,
"grad_norm": 2.4390523433685303,
"learning_rate": 8.06e-06,
"loss": 0.1696,
"step": 1375
},
{
"epoch": 0.28,
"grad_norm": 2.6690866947174072,
"learning_rate": 8.004444444444445e-06,
"loss": 0.1502,
"step": 1400
},
{
"epoch": 0.285,
"grad_norm": 2.090963125228882,
"learning_rate": 7.948888888888889e-06,
"loss": 0.1478,
"step": 1425
},
{
"epoch": 0.29,
"grad_norm": 2.3862431049346924,
"learning_rate": 7.893333333333335e-06,
"loss": 0.1563,
"step": 1450
},
{
"epoch": 0.295,
"grad_norm": 2.7396063804626465,
"learning_rate": 7.837777777777779e-06,
"loss": 0.1589,
"step": 1475
},
{
"epoch": 0.3,
"grad_norm": 2.534780979156494,
"learning_rate": 7.782222222222223e-06,
"loss": 0.1488,
"step": 1500
},
{
"epoch": 0.305,
"grad_norm": 2.8536627292633057,
"learning_rate": 7.726666666666667e-06,
"loss": 0.1731,
"step": 1525
},
{
"epoch": 0.31,
"grad_norm": 2.810988664627075,
"learning_rate": 7.67111111111111e-06,
"loss": 0.1465,
"step": 1550
},
{
"epoch": 0.315,
"grad_norm": 3.3764379024505615,
"learning_rate": 7.6155555555555564e-06,
"loss": 0.2259,
"step": 1575
},
{
"epoch": 0.32,
"grad_norm": 3.14015793800354,
"learning_rate": 7.5600000000000005e-06,
"loss": 0.238,
"step": 1600
},
{
"epoch": 0.325,
"grad_norm": 2.997870683670044,
"learning_rate": 7.504444444444445e-06,
"loss": 0.2381,
"step": 1625
},
{
"epoch": 0.33,
"grad_norm": 2.7839579582214355,
"learning_rate": 7.44888888888889e-06,
"loss": 0.2352,
"step": 1650
},
{
"epoch": 0.335,
"grad_norm": 2.8877196311950684,
"learning_rate": 7.393333333333333e-06,
"loss": 0.2431,
"step": 1675
},
{
"epoch": 0.34,
"grad_norm": 3.2856152057647705,
"learning_rate": 7.337777777777778e-06,
"loss": 0.2279,
"step": 1700
},
{
"epoch": 0.345,
"grad_norm": 3.3398728370666504,
"learning_rate": 7.282222222222222e-06,
"loss": 0.2156,
"step": 1725
},
{
"epoch": 0.35,
"grad_norm": 2.628492593765259,
"learning_rate": 7.226666666666667e-06,
"loss": 0.1938,
"step": 1750
},
{
"epoch": 0.355,
"grad_norm": 3.0386624336242676,
"learning_rate": 7.171111111111112e-06,
"loss": 0.1899,
"step": 1775
},
{
"epoch": 0.36,
"grad_norm": 2.6567234992980957,
"learning_rate": 7.115555555555557e-06,
"loss": 0.1997,
"step": 1800
},
{
"epoch": 0.365,
"grad_norm": 2.5155370235443115,
"learning_rate": 7.06e-06,
"loss": 0.1518,
"step": 1825
},
{
"epoch": 0.37,
"grad_norm": 2.6725313663482666,
"learning_rate": 7.004444444444445e-06,
"loss": 0.1554,
"step": 1850
},
{
"epoch": 0.375,
"grad_norm": 2.657243013381958,
"learning_rate": 6.948888888888889e-06,
"loss": 0.1529,
"step": 1875
},
{
"epoch": 0.38,
"grad_norm": 2.400428295135498,
"learning_rate": 6.893333333333334e-06,
"loss": 0.1556,
"step": 1900
},
{
"epoch": 0.385,
"grad_norm": 2.190220594406128,
"learning_rate": 6.837777777777779e-06,
"loss": 0.1417,
"step": 1925
},
{
"epoch": 0.39,
"grad_norm": 2.150498867034912,
"learning_rate": 6.782222222222222e-06,
"loss": 0.1512,
"step": 1950
},
{
"epoch": 0.395,
"grad_norm": 2.2915945053100586,
"learning_rate": 6.726666666666667e-06,
"loss": 0.1559,
"step": 1975
},
{
"epoch": 0.4,
"grad_norm": 2.43928599357605,
"learning_rate": 6.671111111111112e-06,
"loss": 0.1584,
"step": 2000
},
{
"epoch": 0.4,
"eval_loss": 0.1553979516029358,
"eval_runtime": 1688.9101,
"eval_samples_per_second": 8.973,
"eval_steps_per_second": 1.122,
"eval_wer": 8.175681471754636,
"step": 2000
},
{
"epoch": 0.405,
"grad_norm": 2.3533153533935547,
"learning_rate": 6.615555555555556e-06,
"loss": 0.1576,
"step": 2025
},
{
"epoch": 0.41,
"grad_norm": 2.9315130710601807,
"learning_rate": 6.560000000000001e-06,
"loss": 0.1551,
"step": 2050
},
{
"epoch": 0.415,
"grad_norm": 2.920414924621582,
"learning_rate": 6.504444444444446e-06,
"loss": 0.153,
"step": 2075
},
{
"epoch": 0.42,
"grad_norm": 1.833433985710144,
"learning_rate": 6.448888888888889e-06,
"loss": 0.1299,
"step": 2100
},
{
"epoch": 0.425,
"grad_norm": 2.8404743671417236,
"learning_rate": 6.393333333333334e-06,
"loss": 0.1449,
"step": 2125
},
{
"epoch": 0.43,
"grad_norm": 2.2315993309020996,
"learning_rate": 6.3377777777777786e-06,
"loss": 0.1469,
"step": 2150
},
{
"epoch": 0.435,
"grad_norm": 3.2478203773498535,
"learning_rate": 6.282222222222223e-06,
"loss": 0.1611,
"step": 2175
},
{
"epoch": 0.44,
"grad_norm": 2.173109292984009,
"learning_rate": 6.2266666666666675e-06,
"loss": 0.1383,
"step": 2200
},
{
"epoch": 0.445,
"grad_norm": 2.417632818222046,
"learning_rate": 6.171111111111112e-06,
"loss": 0.1473,
"step": 2225
},
{
"epoch": 0.45,
"grad_norm": 2.384589672088623,
"learning_rate": 6.1155555555555555e-06,
"loss": 0.128,
"step": 2250
},
{
"epoch": 0.455,
"grad_norm": 2.1940202713012695,
"learning_rate": 6.0600000000000004e-06,
"loss": 0.1494,
"step": 2275
},
{
"epoch": 0.46,
"grad_norm": 1.8651658296585083,
"learning_rate": 6.004444444444445e-06,
"loss": 0.1411,
"step": 2300
},
{
"epoch": 0.465,
"grad_norm": 2.0271828174591064,
"learning_rate": 5.948888888888889e-06,
"loss": 0.1373,
"step": 2325
},
{
"epoch": 0.47,
"grad_norm": 2.5282301902770996,
"learning_rate": 5.893333333333334e-06,
"loss": 0.1382,
"step": 2350
},
{
"epoch": 0.475,
"grad_norm": 2.3264193534851074,
"learning_rate": 5.837777777777777e-06,
"loss": 0.1297,
"step": 2375
},
{
"epoch": 0.48,
"grad_norm": 2.045041561126709,
"learning_rate": 5.782222222222222e-06,
"loss": 0.1337,
"step": 2400
},
{
"epoch": 0.485,
"grad_norm": 1.7588191032409668,
"learning_rate": 5.726666666666667e-06,
"loss": 0.1274,
"step": 2425
},
{
"epoch": 0.49,
"grad_norm": 2.436593770980835,
"learning_rate": 5.671111111111112e-06,
"loss": 0.1198,
"step": 2450
},
{
"epoch": 0.495,
"grad_norm": 1.9072929620742798,
"learning_rate": 5.615555555555556e-06,
"loss": 0.1228,
"step": 2475
},
{
"epoch": 0.5,
"grad_norm": 2.490530490875244,
"learning_rate": 5.560000000000001e-06,
"loss": 0.1137,
"step": 2500
},
{
"epoch": 0.505,
"grad_norm": 2.258913516998291,
"learning_rate": 5.504444444444444e-06,
"loss": 0.1389,
"step": 2525
},
{
"epoch": 0.51,
"grad_norm": 2.099606990814209,
"learning_rate": 5.448888888888889e-06,
"loss": 0.1396,
"step": 2550
},
{
"epoch": 0.515,
"grad_norm": 2.8082237243652344,
"learning_rate": 5.393333333333334e-06,
"loss": 0.1709,
"step": 2575
},
{
"epoch": 0.52,
"grad_norm": 2.9639875888824463,
"learning_rate": 5.337777777777779e-06,
"loss": 0.1857,
"step": 2600
},
{
"epoch": 0.525,
"grad_norm": 2.2722744941711426,
"learning_rate": 5.282222222222223e-06,
"loss": 0.1761,
"step": 2625
},
{
"epoch": 0.53,
"grad_norm": 3.0935170650482178,
"learning_rate": 5.226666666666667e-06,
"loss": 0.1566,
"step": 2650
},
{
"epoch": 0.535,
"grad_norm": 1.6952731609344482,
"learning_rate": 5.171111111111111e-06,
"loss": 0.1426,
"step": 2675
},
{
"epoch": 0.54,
"grad_norm": 2.025660991668701,
"learning_rate": 5.115555555555556e-06,
"loss": 0.1525,
"step": 2700
},
{
"epoch": 0.545,
"grad_norm": 2.3472840785980225,
"learning_rate": 5.060000000000001e-06,
"loss": 0.1468,
"step": 2725
},
{
"epoch": 0.55,
"grad_norm": 1.9560490846633911,
"learning_rate": 5.004444444444445e-06,
"loss": 0.1391,
"step": 2750
},
{
"epoch": 0.555,
"grad_norm": 2.2275073528289795,
"learning_rate": 4.94888888888889e-06,
"loss": 0.1444,
"step": 2775
},
{
"epoch": 0.56,
"grad_norm": 2.6753711700439453,
"learning_rate": 4.893333333333334e-06,
"loss": 0.1474,
"step": 2800
},
{
"epoch": 0.565,
"grad_norm": 2.616377353668213,
"learning_rate": 4.837777777777778e-06,
"loss": 0.1266,
"step": 2825
},
{
"epoch": 0.57,
"grad_norm": 2.25164532661438,
"learning_rate": 4.7822222222222226e-06,
"loss": 0.1167,
"step": 2850
},
{
"epoch": 0.575,
"grad_norm": 2.1922624111175537,
"learning_rate": 4.7266666666666674e-06,
"loss": 0.107,
"step": 2875
},
{
"epoch": 0.58,
"grad_norm": 1.9851558208465576,
"learning_rate": 4.6711111111111115e-06,
"loss": 0.1185,
"step": 2900
},
{
"epoch": 0.585,
"grad_norm": 2.494908571243286,
"learning_rate": 4.6155555555555555e-06,
"loss": 0.121,
"step": 2925
},
{
"epoch": 0.59,
"grad_norm": 1.9508136510849,
"learning_rate": 4.56e-06,
"loss": 0.1134,
"step": 2950
},
{
"epoch": 0.595,
"grad_norm": 2.1634881496429443,
"learning_rate": 4.504444444444444e-06,
"loss": 0.1155,
"step": 2975
},
{
"epoch": 0.6,
"grad_norm": 2.3343162536621094,
"learning_rate": 4.448888888888889e-06,
"loss": 0.1309,
"step": 3000
},
{
"epoch": 0.6,
"eval_loss": 0.1425519436597824,
"eval_runtime": 1689.609,
"eval_samples_per_second": 8.97,
"eval_steps_per_second": 1.122,
"eval_wer": 7.414201534837376,
"step": 3000
},
{
"epoch": 0.605,
"grad_norm": 1.8268561363220215,
"learning_rate": 4.393333333333334e-06,
"loss": 0.0938,
"step": 3025
},
{
"epoch": 0.61,
"grad_norm": 2.4155433177948,
"learning_rate": 4.337777777777778e-06,
"loss": 0.1574,
"step": 3050
},
{
"epoch": 0.615,
"grad_norm": 2.5877668857574463,
"learning_rate": 4.282222222222222e-06,
"loss": 0.165,
"step": 3075
},
{
"epoch": 0.62,
"grad_norm": 2.576686382293701,
"learning_rate": 4.226666666666667e-06,
"loss": 0.1851,
"step": 3100
},
{
"epoch": 0.625,
"grad_norm": 3.1633784770965576,
"learning_rate": 4.171111111111111e-06,
"loss": 0.1663,
"step": 3125
},
{
"epoch": 0.63,
"grad_norm": 2.219902753829956,
"learning_rate": 4.115555555555556e-06,
"loss": 0.1451,
"step": 3150
},
{
"epoch": 0.635,
"grad_norm": 1.9273529052734375,
"learning_rate": 4.060000000000001e-06,
"loss": 0.1177,
"step": 3175
},
{
"epoch": 0.64,
"grad_norm": 2.305210590362549,
"learning_rate": 4.004444444444445e-06,
"loss": 0.1125,
"step": 3200
},
{
"epoch": 0.645,
"grad_norm": 2.5342929363250732,
"learning_rate": 3.948888888888889e-06,
"loss": 0.1072,
"step": 3225
},
{
"epoch": 0.65,
"grad_norm": 2.423017978668213,
"learning_rate": 3.893333333333333e-06,
"loss": 0.1262,
"step": 3250
},
{
"epoch": 0.655,
"grad_norm": 2.373019218444824,
"learning_rate": 3.837777777777778e-06,
"loss": 0.118,
"step": 3275
},
{
"epoch": 0.66,
"grad_norm": 1.5576426982879639,
"learning_rate": 3.782222222222223e-06,
"loss": 0.097,
"step": 3300
},
{
"epoch": 0.665,
"grad_norm": 1.6943529844284058,
"learning_rate": 3.726666666666667e-06,
"loss": 0.1067,
"step": 3325
},
{
"epoch": 0.67,
"grad_norm": 1.728334903717041,
"learning_rate": 3.6711111111111113e-06,
"loss": 0.1024,
"step": 3350
},
{
"epoch": 0.675,
"grad_norm": 2.135171890258789,
"learning_rate": 3.615555555555556e-06,
"loss": 0.1093,
"step": 3375
},
{
"epoch": 0.68,
"grad_norm": 1.618424415588379,
"learning_rate": 3.5600000000000002e-06,
"loss": 0.1017,
"step": 3400
},
{
"epoch": 0.685,
"grad_norm": 1.7630062103271484,
"learning_rate": 3.5044444444444447e-06,
"loss": 0.1177,
"step": 3425
},
{
"epoch": 0.69,
"grad_norm": 3.093877077102661,
"learning_rate": 3.4488888888888896e-06,
"loss": 0.1212,
"step": 3450
},
{
"epoch": 0.695,
"grad_norm": 2.9288673400878906,
"learning_rate": 3.3933333333333336e-06,
"loss": 0.1296,
"step": 3475
},
{
"epoch": 0.7,
"grad_norm": 2.8730647563934326,
"learning_rate": 3.337777777777778e-06,
"loss": 0.1247,
"step": 3500
},
{
"epoch": 0.705,
"grad_norm": 2.404902696609497,
"learning_rate": 3.282222222222223e-06,
"loss": 0.118,
"step": 3525
},
{
"epoch": 0.71,
"grad_norm": 2.426954984664917,
"learning_rate": 3.226666666666667e-06,
"loss": 0.1381,
"step": 3550
},
{
"epoch": 0.715,
"grad_norm": 2.6250641345977783,
"learning_rate": 3.1711111111111114e-06,
"loss": 0.1375,
"step": 3575
},
{
"epoch": 0.72,
"grad_norm": 2.184222459793091,
"learning_rate": 3.1155555555555555e-06,
"loss": 0.1227,
"step": 3600
},
{
"epoch": 0.725,
"grad_norm": 2.2613589763641357,
"learning_rate": 3.0600000000000003e-06,
"loss": 0.1225,
"step": 3625
},
{
"epoch": 0.73,
"grad_norm": 2.7714171409606934,
"learning_rate": 3.004444444444445e-06,
"loss": 0.1263,
"step": 3650
},
{
"epoch": 0.735,
"grad_norm": 1.5858285427093506,
"learning_rate": 2.948888888888889e-06,
"loss": 0.1101,
"step": 3675
},
{
"epoch": 0.74,
"grad_norm": 2.5831713676452637,
"learning_rate": 2.8933333333333337e-06,
"loss": 0.0983,
"step": 3700
},
{
"epoch": 0.745,
"grad_norm": 2.260019302368164,
"learning_rate": 2.837777777777778e-06,
"loss": 0.1016,
"step": 3725
},
{
"epoch": 0.75,
"grad_norm": 1.7528841495513916,
"learning_rate": 2.7822222222222222e-06,
"loss": 0.0913,
"step": 3750
},
{
"epoch": 0.755,
"grad_norm": 2.207277536392212,
"learning_rate": 2.726666666666667e-06,
"loss": 0.106,
"step": 3775
},
{
"epoch": 0.76,
"grad_norm": 2.2939164638519287,
"learning_rate": 2.6711111111111116e-06,
"loss": 0.1077,
"step": 3800
},
{
"epoch": 0.765,
"grad_norm": 3.1979310512542725,
"learning_rate": 2.6155555555555556e-06,
"loss": 0.1117,
"step": 3825
},
{
"epoch": 0.77,
"grad_norm": 2.3320703506469727,
"learning_rate": 2.56e-06,
"loss": 0.1517,
"step": 3850
},
{
"epoch": 0.775,
"grad_norm": 2.1557302474975586,
"learning_rate": 2.504444444444445e-06,
"loss": 0.1413,
"step": 3875
},
{
"epoch": 0.78,
"grad_norm": 2.062872886657715,
"learning_rate": 2.448888888888889e-06,
"loss": 0.131,
"step": 3900
},
{
"epoch": 0.785,
"grad_norm": 1.8930553197860718,
"learning_rate": 2.3933333333333334e-06,
"loss": 0.1329,
"step": 3925
},
{
"epoch": 0.79,
"grad_norm": 2.496727228164673,
"learning_rate": 2.337777777777778e-06,
"loss": 0.1188,
"step": 3950
},
{
"epoch": 0.795,
"grad_norm": 2.155111312866211,
"learning_rate": 2.2822222222222223e-06,
"loss": 0.0917,
"step": 3975
},
{
"epoch": 0.8,
"grad_norm": 2.717517137527466,
"learning_rate": 2.226666666666667e-06,
"loss": 0.0984,
"step": 4000
},
{
"epoch": 0.8,
"eval_loss": 0.13701671361923218,
"eval_runtime": 1725.4047,
"eval_samples_per_second": 8.783,
"eval_steps_per_second": 1.098,
"eval_wer": 7.129801315763426,
"step": 4000
},
{
"epoch": 0.805,
"grad_norm": 1.4921413660049438,
"learning_rate": 2.1711111111111113e-06,
"loss": 0.0898,
"step": 4025
},
{
"epoch": 0.81,
"grad_norm": 2.083139657974243,
"learning_rate": 2.1155555555555557e-06,
"loss": 0.0922,
"step": 4050
},
{
"epoch": 0.815,
"grad_norm": 2.0208919048309326,
"learning_rate": 2.06e-06,
"loss": 0.0958,
"step": 4075
},
{
"epoch": 0.82,
"grad_norm": 2.9720261096954346,
"learning_rate": 2.0044444444444446e-06,
"loss": 0.1046,
"step": 4100
},
{
"epoch": 0.825,
"grad_norm": 2.5535240173339844,
"learning_rate": 1.948888888888889e-06,
"loss": 0.1217,
"step": 4125
},
{
"epoch": 0.83,
"grad_norm": 2.354639768600464,
"learning_rate": 1.8933333333333333e-06,
"loss": 0.1314,
"step": 4150
},
{
"epoch": 0.835,
"grad_norm": 2.3064558506011963,
"learning_rate": 1.837777777777778e-06,
"loss": 0.1135,
"step": 4175
},
{
"epoch": 0.84,
"grad_norm": 3.168377161026001,
"learning_rate": 1.7822222222222225e-06,
"loss": 0.1533,
"step": 4200
},
{
"epoch": 0.845,
"grad_norm": 1.9521429538726807,
"learning_rate": 1.728888888888889e-06,
"loss": 0.2723,
"step": 4225
},
{
"epoch": 0.85,
"grad_norm": 2.471975326538086,
"learning_rate": 1.6733333333333335e-06,
"loss": 0.2678,
"step": 4250
},
{
"epoch": 0.855,
"grad_norm": 2.516317367553711,
"learning_rate": 1.6177777777777778e-06,
"loss": 0.2768,
"step": 4275
},
{
"epoch": 0.86,
"grad_norm": 2.5380780696868896,
"learning_rate": 1.5622222222222225e-06,
"loss": 0.2323,
"step": 4300
},
{
"epoch": 0.865,
"grad_norm": 1.7794257402420044,
"learning_rate": 1.506666666666667e-06,
"loss": 0.2052,
"step": 4325
},
{
"epoch": 0.87,
"grad_norm": 2.1327309608459473,
"learning_rate": 1.4511111111111112e-06,
"loss": 0.1311,
"step": 4350
},
{
"epoch": 0.875,
"grad_norm": 1.5396376848220825,
"learning_rate": 1.3955555555555556e-06,
"loss": 0.087,
"step": 4375
},
{
"epoch": 0.88,
"grad_norm": 2.444408416748047,
"learning_rate": 1.34e-06,
"loss": 0.1117,
"step": 4400
},
{
"epoch": 0.885,
"grad_norm": 1.7176605463027954,
"learning_rate": 1.2844444444444445e-06,
"loss": 0.0818,
"step": 4425
},
{
"epoch": 0.89,
"grad_norm": 1.6177054643630981,
"learning_rate": 1.228888888888889e-06,
"loss": 0.0777,
"step": 4450
},
{
"epoch": 0.895,
"grad_norm": 2.648822069168091,
"learning_rate": 1.1733333333333335e-06,
"loss": 0.0901,
"step": 4475
},
{
"epoch": 0.9,
"grad_norm": 1.8845601081848145,
"learning_rate": 1.117777777777778e-06,
"loss": 0.0755,
"step": 4500
},
{
"epoch": 0.905,
"grad_norm": 2.268307685852051,
"learning_rate": 1.0622222222222222e-06,
"loss": 0.0878,
"step": 4525
},
{
"epoch": 0.91,
"grad_norm": 1.6278784275054932,
"learning_rate": 1.0066666666666668e-06,
"loss": 0.1034,
"step": 4550
},
{
"epoch": 0.915,
"grad_norm": 1.8028929233551025,
"learning_rate": 9.511111111111111e-07,
"loss": 0.0883,
"step": 4575
},
{
"epoch": 0.92,
"grad_norm": 1.7106696367263794,
"learning_rate": 8.955555555555557e-07,
"loss": 0.0709,
"step": 4600
},
{
"epoch": 0.925,
"grad_norm": 1.5025192499160767,
"learning_rate": 8.400000000000001e-07,
"loss": 0.0815,
"step": 4625
},
{
"epoch": 0.93,
"grad_norm": 1.554764747619629,
"learning_rate": 7.844444444444445e-07,
"loss": 0.1045,
"step": 4650
},
{
"epoch": 0.935,
"grad_norm": 2.26786732673645,
"learning_rate": 7.28888888888889e-07,
"loss": 0.1028,
"step": 4675
},
{
"epoch": 0.94,
"grad_norm": 2.1246206760406494,
"learning_rate": 6.733333333333334e-07,
"loss": 0.11,
"step": 4700
},
{
"epoch": 0.945,
"grad_norm": 2.038825750350952,
"learning_rate": 6.177777777777778e-07,
"loss": 0.1106,
"step": 4725
},
{
"epoch": 0.95,
"grad_norm": 2.3029303550720215,
"learning_rate": 5.622222222222223e-07,
"loss": 0.1064,
"step": 4750
},
{
"epoch": 0.955,
"grad_norm": 1.985775113105774,
"learning_rate": 5.066666666666667e-07,
"loss": 0.0967,
"step": 4775
},
{
"epoch": 0.96,
"grad_norm": 1.736402988433838,
"learning_rate": 4.511111111111111e-07,
"loss": 0.1028,
"step": 4800
},
{
"epoch": 0.965,
"grad_norm": 2.5676920413970947,
"learning_rate": 3.9555555555555557e-07,
"loss": 0.1014,
"step": 4825
},
{
"epoch": 0.97,
"grad_norm": 2.4653942584991455,
"learning_rate": 3.4000000000000003e-07,
"loss": 0.1041,
"step": 4850
},
{
"epoch": 0.975,
"grad_norm": 2.217700242996216,
"learning_rate": 2.844444444444445e-07,
"loss": 0.1098,
"step": 4875
},
{
"epoch": 0.98,
"grad_norm": 1.7745823860168457,
"learning_rate": 2.2888888888888892e-07,
"loss": 0.1013,
"step": 4900
},
{
"epoch": 0.985,
"grad_norm": 1.8243005275726318,
"learning_rate": 1.7333333333333335e-07,
"loss": 0.1014,
"step": 4925
},
{
"epoch": 0.99,
"grad_norm": 2.028550386428833,
"learning_rate": 1.1777777777777778e-07,
"loss": 0.0992,
"step": 4950
},
{
"epoch": 0.995,
"grad_norm": 1.5399481058120728,
"learning_rate": 6.222222222222223e-08,
"loss": 0.0927,
"step": 4975
},
{
"epoch": 1.0,
"grad_norm": 1.7385917901992798,
"learning_rate": 6.666666666666667e-09,
"loss": 0.0933,
"step": 5000
},
{
"epoch": 1.0,
"eval_loss": 0.13182643055915833,
"eval_runtime": 1689.9204,
"eval_samples_per_second": 8.968,
"eval_steps_per_second": 1.121,
"eval_wer": 6.840122206312234,
"step": 5000
},
{
"epoch": 1.0,
"step": 5000,
"total_flos": 1.632967852032e+20,
"train_loss": 0.15716482088565825,
"train_runtime": 22600.0372,
"train_samples_per_second": 7.08,
"train_steps_per_second": 0.221
}
],
"logging_steps": 25,
"max_steps": 5000,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.632967852032e+20,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}