DewiBrynJones's picture
End of training
04988e0 verified
raw
history blame contribute delete
No virus
36.7 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 9.727626459143968,
"eval_steps": 1000,
"global_step": 5000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.048638132295719845,
"grad_norm": 7.16464900970459,
"learning_rate": 5.000000000000001e-07,
"loss": 1.0297,
"step": 25
},
{
"epoch": 0.09727626459143969,
"grad_norm": 4.811949729919434,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.786,
"step": 50
},
{
"epoch": 0.14591439688715954,
"grad_norm": 4.57177734375,
"learning_rate": 1.5e-06,
"loss": 0.5046,
"step": 75
},
{
"epoch": 0.19455252918287938,
"grad_norm": 4.517965793609619,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.4671,
"step": 100
},
{
"epoch": 0.24319066147859922,
"grad_norm": 4.065127849578857,
"learning_rate": 2.5e-06,
"loss": 0.4381,
"step": 125
},
{
"epoch": 0.2918287937743191,
"grad_norm": 3.5869154930114746,
"learning_rate": 3e-06,
"loss": 0.4062,
"step": 150
},
{
"epoch": 0.3404669260700389,
"grad_norm": 4.262431621551514,
"learning_rate": 3.5e-06,
"loss": 0.3951,
"step": 175
},
{
"epoch": 0.38910505836575876,
"grad_norm": 4.2176713943481445,
"learning_rate": 4.000000000000001e-06,
"loss": 0.3862,
"step": 200
},
{
"epoch": 0.4377431906614786,
"grad_norm": 3.8358681201934814,
"learning_rate": 4.5e-06,
"loss": 0.364,
"step": 225
},
{
"epoch": 0.48638132295719844,
"grad_norm": 3.8311147689819336,
"learning_rate": 5e-06,
"loss": 0.3425,
"step": 250
},
{
"epoch": 0.5350194552529183,
"grad_norm": 3.7137532234191895,
"learning_rate": 5.500000000000001e-06,
"loss": 0.339,
"step": 275
},
{
"epoch": 0.5836575875486382,
"grad_norm": 4.2145280838012695,
"learning_rate": 6e-06,
"loss": 0.3393,
"step": 300
},
{
"epoch": 0.632295719844358,
"grad_norm": 3.540754556655884,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.3154,
"step": 325
},
{
"epoch": 0.6809338521400778,
"grad_norm": 3.7712759971618652,
"learning_rate": 7e-06,
"loss": 0.319,
"step": 350
},
{
"epoch": 0.7295719844357976,
"grad_norm": 3.6147518157958984,
"learning_rate": 7.500000000000001e-06,
"loss": 0.3127,
"step": 375
},
{
"epoch": 0.7782101167315175,
"grad_norm": 3.6948869228363037,
"learning_rate": 8.000000000000001e-06,
"loss": 0.2965,
"step": 400
},
{
"epoch": 0.8268482490272373,
"grad_norm": 3.350135087966919,
"learning_rate": 8.5e-06,
"loss": 0.3001,
"step": 425
},
{
"epoch": 0.8754863813229572,
"grad_norm": 3.7692646980285645,
"learning_rate": 9e-06,
"loss": 0.281,
"step": 450
},
{
"epoch": 0.9241245136186771,
"grad_norm": 3.2747561931610107,
"learning_rate": 9.5e-06,
"loss": 0.2707,
"step": 475
},
{
"epoch": 0.9727626459143969,
"grad_norm": 3.260094165802002,
"learning_rate": 1e-05,
"loss": 0.2708,
"step": 500
},
{
"epoch": 1.0214007782101167,
"grad_norm": 2.917370080947876,
"learning_rate": 9.944444444444445e-06,
"loss": 0.2415,
"step": 525
},
{
"epoch": 1.0700389105058365,
"grad_norm": 2.564760208129883,
"learning_rate": 9.88888888888889e-06,
"loss": 0.1714,
"step": 550
},
{
"epoch": 1.1186770428015564,
"grad_norm": 2.844514846801758,
"learning_rate": 9.833333333333333e-06,
"loss": 0.1842,
"step": 575
},
{
"epoch": 1.1673151750972763,
"grad_norm": 3.1123480796813965,
"learning_rate": 9.777777777777779e-06,
"loss": 0.171,
"step": 600
},
{
"epoch": 1.2159533073929962,
"grad_norm": 3.253563642501831,
"learning_rate": 9.722222222222223e-06,
"loss": 0.1773,
"step": 625
},
{
"epoch": 1.264591439688716,
"grad_norm": 2.3145253658294678,
"learning_rate": 9.666666666666667e-06,
"loss": 0.1779,
"step": 650
},
{
"epoch": 1.3132295719844358,
"grad_norm": 2.493607521057129,
"learning_rate": 9.611111111111112e-06,
"loss": 0.1717,
"step": 675
},
{
"epoch": 1.3618677042801557,
"grad_norm": 2.081571578979492,
"learning_rate": 9.555555555555556e-06,
"loss": 0.1747,
"step": 700
},
{
"epoch": 1.4105058365758754,
"grad_norm": 2.9359965324401855,
"learning_rate": 9.5e-06,
"loss": 0.1591,
"step": 725
},
{
"epoch": 1.4591439688715953,
"grad_norm": 2.9549202919006348,
"learning_rate": 9.444444444444445e-06,
"loss": 0.1556,
"step": 750
},
{
"epoch": 1.5077821011673151,
"grad_norm": 3.0995261669158936,
"learning_rate": 9.38888888888889e-06,
"loss": 0.1553,
"step": 775
},
{
"epoch": 1.556420233463035,
"grad_norm": 2.3519606590270996,
"learning_rate": 9.333333333333334e-06,
"loss": 0.1502,
"step": 800
},
{
"epoch": 1.605058365758755,
"grad_norm": 2.3519246578216553,
"learning_rate": 9.277777777777778e-06,
"loss": 0.1398,
"step": 825
},
{
"epoch": 1.6536964980544746,
"grad_norm": 2.292020797729492,
"learning_rate": 9.222222222222224e-06,
"loss": 0.1518,
"step": 850
},
{
"epoch": 1.7023346303501945,
"grad_norm": 2.7264564037323,
"learning_rate": 9.166666666666666e-06,
"loss": 0.1393,
"step": 875
},
{
"epoch": 1.7509727626459144,
"grad_norm": 2.202409267425537,
"learning_rate": 9.111111111111112e-06,
"loss": 0.1432,
"step": 900
},
{
"epoch": 1.7996108949416343,
"grad_norm": 2.6967380046844482,
"learning_rate": 9.055555555555556e-06,
"loss": 0.1492,
"step": 925
},
{
"epoch": 1.8482490272373542,
"grad_norm": 2.8609344959259033,
"learning_rate": 9e-06,
"loss": 0.1351,
"step": 950
},
{
"epoch": 1.8968871595330739,
"grad_norm": 1.9660629034042358,
"learning_rate": 8.944444444444446e-06,
"loss": 0.1378,
"step": 975
},
{
"epoch": 1.9455252918287937,
"grad_norm": 2.9408161640167236,
"learning_rate": 8.888888888888888e-06,
"loss": 0.1429,
"step": 1000
},
{
"epoch": 1.9455252918287937,
"eval_loss": 0.2754097580909729,
"eval_runtime": 2154.7003,
"eval_samples_per_second": 2.497,
"eval_steps_per_second": 0.156,
"eval_wer": 0.22082388961787744,
"step": 1000
},
{
"epoch": 1.9941634241245136,
"grad_norm": 2.337907075881958,
"learning_rate": 8.833333333333334e-06,
"loss": 0.137,
"step": 1025
},
{
"epoch": 2.0428015564202333,
"grad_norm": 1.9045900106430054,
"learning_rate": 8.777777777777778e-06,
"loss": 0.0728,
"step": 1050
},
{
"epoch": 2.0914396887159534,
"grad_norm": 2.0690503120422363,
"learning_rate": 8.722222222222224e-06,
"loss": 0.0615,
"step": 1075
},
{
"epoch": 2.140077821011673,
"grad_norm": 2.0722174644470215,
"learning_rate": 8.666666666666668e-06,
"loss": 0.0582,
"step": 1100
},
{
"epoch": 2.188715953307393,
"grad_norm": 1.7977361679077148,
"learning_rate": 8.611111111111112e-06,
"loss": 0.0589,
"step": 1125
},
{
"epoch": 2.237354085603113,
"grad_norm": 1.5375025272369385,
"learning_rate": 8.555555555555556e-06,
"loss": 0.0657,
"step": 1150
},
{
"epoch": 2.2859922178988326,
"grad_norm": 1.9439505338668823,
"learning_rate": 8.5e-06,
"loss": 0.0648,
"step": 1175
},
{
"epoch": 2.3346303501945527,
"grad_norm": 1.7909560203552246,
"learning_rate": 8.444444444444446e-06,
"loss": 0.0594,
"step": 1200
},
{
"epoch": 2.3832684824902723,
"grad_norm": 1.7498406171798706,
"learning_rate": 8.38888888888889e-06,
"loss": 0.0627,
"step": 1225
},
{
"epoch": 2.4319066147859925,
"grad_norm": 1.557906985282898,
"learning_rate": 8.333333333333334e-06,
"loss": 0.0585,
"step": 1250
},
{
"epoch": 2.480544747081712,
"grad_norm": 1.3372323513031006,
"learning_rate": 8.277777777777778e-06,
"loss": 0.0614,
"step": 1275
},
{
"epoch": 2.529182879377432,
"grad_norm": 1.8402326107025146,
"learning_rate": 8.222222222222222e-06,
"loss": 0.0601,
"step": 1300
},
{
"epoch": 2.5778210116731515,
"grad_norm": 2.003889560699463,
"learning_rate": 8.166666666666668e-06,
"loss": 0.0575,
"step": 1325
},
{
"epoch": 2.6264591439688716,
"grad_norm": 1.5208065509796143,
"learning_rate": 8.111111111111112e-06,
"loss": 0.0563,
"step": 1350
},
{
"epoch": 2.6750972762645917,
"grad_norm": 2.1009633541107178,
"learning_rate": 8.055555555555557e-06,
"loss": 0.0633,
"step": 1375
},
{
"epoch": 2.7237354085603114,
"grad_norm": 1.367943286895752,
"learning_rate": 8.000000000000001e-06,
"loss": 0.0572,
"step": 1400
},
{
"epoch": 2.772373540856031,
"grad_norm": 1.914768934249878,
"learning_rate": 7.944444444444445e-06,
"loss": 0.0541,
"step": 1425
},
{
"epoch": 2.8210116731517507,
"grad_norm": 2.1714184284210205,
"learning_rate": 7.88888888888889e-06,
"loss": 0.059,
"step": 1450
},
{
"epoch": 2.869649805447471,
"grad_norm": 1.951029658317566,
"learning_rate": 7.833333333333333e-06,
"loss": 0.0571,
"step": 1475
},
{
"epoch": 2.9182879377431905,
"grad_norm": 1.6378835439682007,
"learning_rate": 7.77777777777778e-06,
"loss": 0.0563,
"step": 1500
},
{
"epoch": 2.9669260700389106,
"grad_norm": 1.8460659980773926,
"learning_rate": 7.722222222222223e-06,
"loss": 0.0556,
"step": 1525
},
{
"epoch": 3.0155642023346303,
"grad_norm": 1.0332473516464233,
"learning_rate": 7.666666666666667e-06,
"loss": 0.045,
"step": 1550
},
{
"epoch": 3.06420233463035,
"grad_norm": 1.5045939683914185,
"learning_rate": 7.611111111111111e-06,
"loss": 0.0227,
"step": 1575
},
{
"epoch": 3.11284046692607,
"grad_norm": 2.1660637855529785,
"learning_rate": 7.555555555555556e-06,
"loss": 0.0275,
"step": 1600
},
{
"epoch": 3.1614785992217898,
"grad_norm": 1.159175992012024,
"learning_rate": 7.500000000000001e-06,
"loss": 0.024,
"step": 1625
},
{
"epoch": 3.21011673151751,
"grad_norm": 0.9240212440490723,
"learning_rate": 7.444444444444445e-06,
"loss": 0.0271,
"step": 1650
},
{
"epoch": 3.2587548638132295,
"grad_norm": 0.9973633885383606,
"learning_rate": 7.38888888888889e-06,
"loss": 0.0266,
"step": 1675
},
{
"epoch": 3.307392996108949,
"grad_norm": 1.5561943054199219,
"learning_rate": 7.333333333333333e-06,
"loss": 0.0267,
"step": 1700
},
{
"epoch": 3.3560311284046693,
"grad_norm": 1.2981663942337036,
"learning_rate": 7.277777777777778e-06,
"loss": 0.0284,
"step": 1725
},
{
"epoch": 3.404669260700389,
"grad_norm": 1.1085954904556274,
"learning_rate": 7.222222222222223e-06,
"loss": 0.0239,
"step": 1750
},
{
"epoch": 3.453307392996109,
"grad_norm": 1.219834566116333,
"learning_rate": 7.166666666666667e-06,
"loss": 0.0244,
"step": 1775
},
{
"epoch": 3.501945525291829,
"grad_norm": 1.1478172540664673,
"learning_rate": 7.111111111111112e-06,
"loss": 0.0273,
"step": 1800
},
{
"epoch": 3.5505836575875485,
"grad_norm": 1.5546625852584839,
"learning_rate": 7.055555555555557e-06,
"loss": 0.0242,
"step": 1825
},
{
"epoch": 3.5992217898832686,
"grad_norm": 1.5303540229797363,
"learning_rate": 7e-06,
"loss": 0.0248,
"step": 1850
},
{
"epoch": 3.6478599221789882,
"grad_norm": 1.2220957279205322,
"learning_rate": 6.944444444444445e-06,
"loss": 0.0268,
"step": 1875
},
{
"epoch": 3.6964980544747084,
"grad_norm": 1.076720952987671,
"learning_rate": 6.88888888888889e-06,
"loss": 0.0246,
"step": 1900
},
{
"epoch": 3.745136186770428,
"grad_norm": 0.6398268342018127,
"learning_rate": 6.833333333333334e-06,
"loss": 0.0251,
"step": 1925
},
{
"epoch": 3.7937743190661477,
"grad_norm": 1.0957626104354858,
"learning_rate": 6.777777777777779e-06,
"loss": 0.024,
"step": 1950
},
{
"epoch": 3.842412451361868,
"grad_norm": 1.3786637783050537,
"learning_rate": 6.7222222222222235e-06,
"loss": 0.0221,
"step": 1975
},
{
"epoch": 3.8910505836575875,
"grad_norm": 1.0472965240478516,
"learning_rate": 6.666666666666667e-06,
"loss": 0.0232,
"step": 2000
},
{
"epoch": 3.8910505836575875,
"eval_loss": 0.2916410565376282,
"eval_runtime": 2132.1371,
"eval_samples_per_second": 2.524,
"eval_steps_per_second": 0.158,
"eval_wer": 0.19905364030378941,
"step": 2000
},
{
"epoch": 3.9396887159533076,
"grad_norm": 1.0177409648895264,
"learning_rate": 6.6111111111111115e-06,
"loss": 0.0238,
"step": 2025
},
{
"epoch": 3.9883268482490273,
"grad_norm": 1.0486767292022705,
"learning_rate": 6.555555555555556e-06,
"loss": 0.0245,
"step": 2050
},
{
"epoch": 4.036964980544747,
"grad_norm": 0.7433611750602722,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.0138,
"step": 2075
},
{
"epoch": 4.085603112840467,
"grad_norm": 0.4771675765514374,
"learning_rate": 6.444444444444445e-06,
"loss": 0.0109,
"step": 2100
},
{
"epoch": 4.134241245136187,
"grad_norm": 0.6232401132583618,
"learning_rate": 6.3888888888888885e-06,
"loss": 0.0109,
"step": 2125
},
{
"epoch": 4.182879377431907,
"grad_norm": 0.6728478670120239,
"learning_rate": 6.333333333333333e-06,
"loss": 0.0103,
"step": 2150
},
{
"epoch": 4.2315175097276265,
"grad_norm": 1.6818970441818237,
"learning_rate": 6.277777777777778e-06,
"loss": 0.0114,
"step": 2175
},
{
"epoch": 4.280155642023346,
"grad_norm": 0.36247119307518005,
"learning_rate": 6.222222222222223e-06,
"loss": 0.011,
"step": 2200
},
{
"epoch": 4.328793774319066,
"grad_norm": 1.3438676595687866,
"learning_rate": 6.166666666666667e-06,
"loss": 0.0126,
"step": 2225
},
{
"epoch": 4.377431906614786,
"grad_norm": 0.8739299774169922,
"learning_rate": 6.111111111111112e-06,
"loss": 0.0104,
"step": 2250
},
{
"epoch": 4.426070038910506,
"grad_norm": 0.6817359328269958,
"learning_rate": 6.055555555555555e-06,
"loss": 0.0118,
"step": 2275
},
{
"epoch": 4.474708171206226,
"grad_norm": 0.5845357775688171,
"learning_rate": 6e-06,
"loss": 0.0125,
"step": 2300
},
{
"epoch": 4.523346303501945,
"grad_norm": 1.067460298538208,
"learning_rate": 5.944444444444445e-06,
"loss": 0.0099,
"step": 2325
},
{
"epoch": 4.571984435797665,
"grad_norm": 0.7725332379341125,
"learning_rate": 5.88888888888889e-06,
"loss": 0.0123,
"step": 2350
},
{
"epoch": 4.620622568093385,
"grad_norm": 0.8168444633483887,
"learning_rate": 5.833333333333334e-06,
"loss": 0.01,
"step": 2375
},
{
"epoch": 4.669260700389105,
"grad_norm": 0.7538411617279053,
"learning_rate": 5.777777777777778e-06,
"loss": 0.0098,
"step": 2400
},
{
"epoch": 4.717898832684825,
"grad_norm": 0.3141120970249176,
"learning_rate": 5.722222222222222e-06,
"loss": 0.0107,
"step": 2425
},
{
"epoch": 4.766536964980545,
"grad_norm": 0.887184202671051,
"learning_rate": 5.666666666666667e-06,
"loss": 0.0095,
"step": 2450
},
{
"epoch": 4.815175097276264,
"grad_norm": 0.5846179127693176,
"learning_rate": 5.611111111111112e-06,
"loss": 0.0099,
"step": 2475
},
{
"epoch": 4.863813229571985,
"grad_norm": 0.3938518464565277,
"learning_rate": 5.555555555555557e-06,
"loss": 0.0091,
"step": 2500
},
{
"epoch": 4.912451361867705,
"grad_norm": 0.9777388572692871,
"learning_rate": 5.500000000000001e-06,
"loss": 0.0093,
"step": 2525
},
{
"epoch": 4.961089494163424,
"grad_norm": 0.7862864136695862,
"learning_rate": 5.444444444444445e-06,
"loss": 0.0111,
"step": 2550
},
{
"epoch": 5.009727626459144,
"grad_norm": 0.7504919171333313,
"learning_rate": 5.388888888888889e-06,
"loss": 0.0094,
"step": 2575
},
{
"epoch": 5.058365758754864,
"grad_norm": 0.2351156324148178,
"learning_rate": 5.333333333333334e-06,
"loss": 0.0049,
"step": 2600
},
{
"epoch": 5.107003891050583,
"grad_norm": 0.13562794029712677,
"learning_rate": 5.2777777777777785e-06,
"loss": 0.0043,
"step": 2625
},
{
"epoch": 5.155642023346304,
"grad_norm": 0.33899495005607605,
"learning_rate": 5.2222222222222226e-06,
"loss": 0.0048,
"step": 2650
},
{
"epoch": 5.2042801556420235,
"grad_norm": 0.9408302307128906,
"learning_rate": 5.1666666666666675e-06,
"loss": 0.0051,
"step": 2675
},
{
"epoch": 5.252918287937743,
"grad_norm": 0.35573306679725647,
"learning_rate": 5.1111111111111115e-06,
"loss": 0.0053,
"step": 2700
},
{
"epoch": 5.301556420233463,
"grad_norm": 0.7676318883895874,
"learning_rate": 5.0555555555555555e-06,
"loss": 0.0042,
"step": 2725
},
{
"epoch": 5.3501945525291825,
"grad_norm": 0.6776612401008606,
"learning_rate": 5e-06,
"loss": 0.004,
"step": 2750
},
{
"epoch": 5.398832684824903,
"grad_norm": 1.0338044166564941,
"learning_rate": 4.944444444444445e-06,
"loss": 0.0047,
"step": 2775
},
{
"epoch": 5.447470817120623,
"grad_norm": 0.2744717001914978,
"learning_rate": 4.888888888888889e-06,
"loss": 0.0068,
"step": 2800
},
{
"epoch": 5.496108949416342,
"grad_norm": 0.7075155973434448,
"learning_rate": 4.833333333333333e-06,
"loss": 0.0061,
"step": 2825
},
{
"epoch": 5.544747081712062,
"grad_norm": 0.1820390671491623,
"learning_rate": 4.777777777777778e-06,
"loss": 0.0052,
"step": 2850
},
{
"epoch": 5.593385214007782,
"grad_norm": 0.33624935150146484,
"learning_rate": 4.722222222222222e-06,
"loss": 0.0038,
"step": 2875
},
{
"epoch": 5.642023346303502,
"grad_norm": 0.1461501270532608,
"learning_rate": 4.666666666666667e-06,
"loss": 0.0046,
"step": 2900
},
{
"epoch": 5.690661478599222,
"grad_norm": 0.33367106318473816,
"learning_rate": 4.611111111111112e-06,
"loss": 0.0042,
"step": 2925
},
{
"epoch": 5.739299610894942,
"grad_norm": 0.32830071449279785,
"learning_rate": 4.555555555555556e-06,
"loss": 0.0043,
"step": 2950
},
{
"epoch": 5.787937743190661,
"grad_norm": 1.2332854270935059,
"learning_rate": 4.5e-06,
"loss": 0.0039,
"step": 2975
},
{
"epoch": 5.836575875486381,
"grad_norm": 0.6161640286445618,
"learning_rate": 4.444444444444444e-06,
"loss": 0.0046,
"step": 3000
},
{
"epoch": 5.836575875486381,
"eval_loss": 0.3218872845172882,
"eval_runtime": 2114.549,
"eval_samples_per_second": 2.545,
"eval_steps_per_second": 0.159,
"eval_wer": 0.18784047079406735,
"step": 3000
},
{
"epoch": 5.885214007782102,
"grad_norm": 2.608947992324829,
"learning_rate": 4.388888888888889e-06,
"loss": 0.0061,
"step": 3025
},
{
"epoch": 5.933852140077821,
"grad_norm": 0.25549063086509705,
"learning_rate": 4.333333333333334e-06,
"loss": 0.0044,
"step": 3050
},
{
"epoch": 5.982490272373541,
"grad_norm": 1.4720717668533325,
"learning_rate": 4.277777777777778e-06,
"loss": 0.0051,
"step": 3075
},
{
"epoch": 6.031128404669261,
"grad_norm": 0.7722542881965637,
"learning_rate": 4.222222222222223e-06,
"loss": 0.0037,
"step": 3100
},
{
"epoch": 6.07976653696498,
"grad_norm": 0.35967299342155457,
"learning_rate": 4.166666666666667e-06,
"loss": 0.0029,
"step": 3125
},
{
"epoch": 6.1284046692607,
"grad_norm": 0.7472084164619446,
"learning_rate": 4.111111111111111e-06,
"loss": 0.0029,
"step": 3150
},
{
"epoch": 6.1770428015564205,
"grad_norm": 0.08338561654090881,
"learning_rate": 4.055555555555556e-06,
"loss": 0.0023,
"step": 3175
},
{
"epoch": 6.22568093385214,
"grad_norm": 0.49297797679901123,
"learning_rate": 4.000000000000001e-06,
"loss": 0.0022,
"step": 3200
},
{
"epoch": 6.27431906614786,
"grad_norm": 0.08119112253189087,
"learning_rate": 3.944444444444445e-06,
"loss": 0.0019,
"step": 3225
},
{
"epoch": 6.3229571984435795,
"grad_norm": 0.13106492161750793,
"learning_rate": 3.88888888888889e-06,
"loss": 0.0027,
"step": 3250
},
{
"epoch": 6.3715953307393,
"grad_norm": 0.18573686480522156,
"learning_rate": 3.833333333333334e-06,
"loss": 0.003,
"step": 3275
},
{
"epoch": 6.42023346303502,
"grad_norm": 0.09228133410215378,
"learning_rate": 3.777777777777778e-06,
"loss": 0.004,
"step": 3300
},
{
"epoch": 6.468871595330739,
"grad_norm": 0.8169698715209961,
"learning_rate": 3.7222222222222225e-06,
"loss": 0.0029,
"step": 3325
},
{
"epoch": 6.517509727626459,
"grad_norm": 0.23130229115486145,
"learning_rate": 3.6666666666666666e-06,
"loss": 0.0026,
"step": 3350
},
{
"epoch": 6.566147859922179,
"grad_norm": 0.10798731446266174,
"learning_rate": 3.6111111111111115e-06,
"loss": 0.0036,
"step": 3375
},
{
"epoch": 6.614785992217898,
"grad_norm": 0.09145894646644592,
"learning_rate": 3.555555555555556e-06,
"loss": 0.0024,
"step": 3400
},
{
"epoch": 6.663424124513619,
"grad_norm": 0.3099443018436432,
"learning_rate": 3.5e-06,
"loss": 0.0018,
"step": 3425
},
{
"epoch": 6.712062256809339,
"grad_norm": 0.13762035965919495,
"learning_rate": 3.444444444444445e-06,
"loss": 0.0021,
"step": 3450
},
{
"epoch": 6.760700389105058,
"grad_norm": 0.11253529787063599,
"learning_rate": 3.3888888888888893e-06,
"loss": 0.0012,
"step": 3475
},
{
"epoch": 6.809338521400778,
"grad_norm": 0.054800793528556824,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.0019,
"step": 3500
},
{
"epoch": 6.857976653696498,
"grad_norm": 0.06803842633962631,
"learning_rate": 3.277777777777778e-06,
"loss": 0.0024,
"step": 3525
},
{
"epoch": 6.906614785992218,
"grad_norm": 0.08114957064390182,
"learning_rate": 3.2222222222222227e-06,
"loss": 0.0019,
"step": 3550
},
{
"epoch": 6.955252918287938,
"grad_norm": 0.06155654788017273,
"learning_rate": 3.1666666666666667e-06,
"loss": 0.0017,
"step": 3575
},
{
"epoch": 7.003891050583658,
"grad_norm": 0.09375619143247604,
"learning_rate": 3.1111111111111116e-06,
"loss": 0.0015,
"step": 3600
},
{
"epoch": 7.052529182879377,
"grad_norm": 0.04578198865056038,
"learning_rate": 3.055555555555556e-06,
"loss": 0.0011,
"step": 3625
},
{
"epoch": 7.101167315175097,
"grad_norm": 0.036920215934515,
"learning_rate": 3e-06,
"loss": 0.0012,
"step": 3650
},
{
"epoch": 7.1498054474708175,
"grad_norm": 0.07471055537462234,
"learning_rate": 2.944444444444445e-06,
"loss": 0.0009,
"step": 3675
},
{
"epoch": 7.198443579766537,
"grad_norm": 0.05370425805449486,
"learning_rate": 2.888888888888889e-06,
"loss": 0.0009,
"step": 3700
},
{
"epoch": 7.247081712062257,
"grad_norm": 0.04795600846409798,
"learning_rate": 2.8333333333333335e-06,
"loss": 0.001,
"step": 3725
},
{
"epoch": 7.2957198443579765,
"grad_norm": 0.030901705846190453,
"learning_rate": 2.7777777777777783e-06,
"loss": 0.0007,
"step": 3750
},
{
"epoch": 7.344357976653696,
"grad_norm": 0.05389130115509033,
"learning_rate": 2.7222222222222224e-06,
"loss": 0.0012,
"step": 3775
},
{
"epoch": 7.392996108949417,
"grad_norm": 0.05646834522485733,
"learning_rate": 2.666666666666667e-06,
"loss": 0.0009,
"step": 3800
},
{
"epoch": 7.441634241245136,
"grad_norm": 0.04085018113255501,
"learning_rate": 2.6111111111111113e-06,
"loss": 0.001,
"step": 3825
},
{
"epoch": 7.490272373540856,
"grad_norm": 0.024638062343001366,
"learning_rate": 2.5555555555555557e-06,
"loss": 0.0009,
"step": 3850
},
{
"epoch": 7.538910505836576,
"grad_norm": 0.03571788966655731,
"learning_rate": 2.5e-06,
"loss": 0.0011,
"step": 3875
},
{
"epoch": 7.587548638132295,
"grad_norm": 0.05909154936671257,
"learning_rate": 2.4444444444444447e-06,
"loss": 0.0007,
"step": 3900
},
{
"epoch": 7.636186770428015,
"grad_norm": 0.028353404253721237,
"learning_rate": 2.388888888888889e-06,
"loss": 0.001,
"step": 3925
},
{
"epoch": 7.684824902723736,
"grad_norm": 0.16029401123523712,
"learning_rate": 2.3333333333333336e-06,
"loss": 0.0009,
"step": 3950
},
{
"epoch": 7.733463035019455,
"grad_norm": 0.04438905417919159,
"learning_rate": 2.277777777777778e-06,
"loss": 0.0008,
"step": 3975
},
{
"epoch": 7.782101167315175,
"grad_norm": 0.035433579236269,
"learning_rate": 2.222222222222222e-06,
"loss": 0.0009,
"step": 4000
},
{
"epoch": 7.782101167315175,
"eval_loss": 0.3454054296016693,
"eval_runtime": 2148.9325,
"eval_samples_per_second": 2.504,
"eval_steps_per_second": 0.157,
"eval_wer": 0.1831881983379061,
"step": 4000
},
{
"epoch": 7.830739299610895,
"grad_norm": 0.028996312990784645,
"learning_rate": 2.166666666666667e-06,
"loss": 0.0008,
"step": 4025
},
{
"epoch": 7.879377431906615,
"grad_norm": 0.038502488285303116,
"learning_rate": 2.1111111111111114e-06,
"loss": 0.0014,
"step": 4050
},
{
"epoch": 7.928015564202335,
"grad_norm": 0.06861083209514618,
"learning_rate": 2.0555555555555555e-06,
"loss": 0.0009,
"step": 4075
},
{
"epoch": 7.976653696498055,
"grad_norm": 0.07425787299871445,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.0011,
"step": 4100
},
{
"epoch": 8.025291828793774,
"grad_norm": 0.025021173059940338,
"learning_rate": 1.944444444444445e-06,
"loss": 0.0006,
"step": 4125
},
{
"epoch": 8.073929961089494,
"grad_norm": 0.031610701233148575,
"learning_rate": 1.888888888888889e-06,
"loss": 0.0006,
"step": 4150
},
{
"epoch": 8.122568093385214,
"grad_norm": 0.02210886962711811,
"learning_rate": 1.8333333333333333e-06,
"loss": 0.0008,
"step": 4175
},
{
"epoch": 8.171206225680933,
"grad_norm": 0.03183664008975029,
"learning_rate": 1.777777777777778e-06,
"loss": 0.0007,
"step": 4200
},
{
"epoch": 8.219844357976653,
"grad_norm": 0.024653365835547447,
"learning_rate": 1.7222222222222224e-06,
"loss": 0.0005,
"step": 4225
},
{
"epoch": 8.268482490272374,
"grad_norm": 0.02229795977473259,
"learning_rate": 1.6666666666666667e-06,
"loss": 0.0005,
"step": 4250
},
{
"epoch": 8.317120622568094,
"grad_norm": 0.025678085163235664,
"learning_rate": 1.6111111111111113e-06,
"loss": 0.0005,
"step": 4275
},
{
"epoch": 8.365758754863814,
"grad_norm": 0.025166384875774384,
"learning_rate": 1.5555555555555558e-06,
"loss": 0.0005,
"step": 4300
},
{
"epoch": 8.414396887159533,
"grad_norm": 0.017260603606700897,
"learning_rate": 1.5e-06,
"loss": 0.0005,
"step": 4325
},
{
"epoch": 8.463035019455253,
"grad_norm": 0.016240952536463737,
"learning_rate": 1.4444444444444445e-06,
"loss": 0.0005,
"step": 4350
},
{
"epoch": 8.511673151750973,
"grad_norm": 0.021515797823667526,
"learning_rate": 1.3888888888888892e-06,
"loss": 0.0006,
"step": 4375
},
{
"epoch": 8.560311284046692,
"grad_norm": 0.023731861263513565,
"learning_rate": 1.3333333333333334e-06,
"loss": 0.0006,
"step": 4400
},
{
"epoch": 8.608949416342412,
"grad_norm": 0.027586709707975388,
"learning_rate": 1.2777777777777779e-06,
"loss": 0.0006,
"step": 4425
},
{
"epoch": 8.657587548638132,
"grad_norm": 0.04710078611969948,
"learning_rate": 1.2222222222222223e-06,
"loss": 0.0005,
"step": 4450
},
{
"epoch": 8.706225680933851,
"grad_norm": 0.030895188450813293,
"learning_rate": 1.1666666666666668e-06,
"loss": 0.0005,
"step": 4475
},
{
"epoch": 8.754863813229573,
"grad_norm": 0.04257024824619293,
"learning_rate": 1.111111111111111e-06,
"loss": 0.0006,
"step": 4500
},
{
"epoch": 8.803501945525293,
"grad_norm": 0.022679895162582397,
"learning_rate": 1.0555555555555557e-06,
"loss": 0.0006,
"step": 4525
},
{
"epoch": 8.852140077821012,
"grad_norm": 0.028052836656570435,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.0006,
"step": 4550
},
{
"epoch": 8.900778210116732,
"grad_norm": 0.022373031824827194,
"learning_rate": 9.444444444444445e-07,
"loss": 0.0005,
"step": 4575
},
{
"epoch": 8.949416342412452,
"grad_norm": 0.02304161712527275,
"learning_rate": 8.88888888888889e-07,
"loss": 0.0005,
"step": 4600
},
{
"epoch": 8.998054474708171,
"grad_norm": 0.020944086834788322,
"learning_rate": 8.333333333333333e-07,
"loss": 0.0005,
"step": 4625
},
{
"epoch": 9.04669260700389,
"grad_norm": 0.018772481009364128,
"learning_rate": 7.777777777777779e-07,
"loss": 0.0005,
"step": 4650
},
{
"epoch": 9.09533073929961,
"grad_norm": 0.019685110077261925,
"learning_rate": 7.222222222222222e-07,
"loss": 0.0005,
"step": 4675
},
{
"epoch": 9.14396887159533,
"grad_norm": 0.017540233209729195,
"learning_rate": 6.666666666666667e-07,
"loss": 0.0004,
"step": 4700
},
{
"epoch": 9.19260700389105,
"grad_norm": 0.022134965285658836,
"learning_rate": 6.111111111111112e-07,
"loss": 0.0004,
"step": 4725
},
{
"epoch": 9.24124513618677,
"grad_norm": 0.01725279912352562,
"learning_rate": 5.555555555555555e-07,
"loss": 0.0005,
"step": 4750
},
{
"epoch": 9.289883268482491,
"grad_norm": 0.01975845918059349,
"learning_rate": 5.000000000000001e-07,
"loss": 0.0004,
"step": 4775
},
{
"epoch": 9.33852140077821,
"grad_norm": 0.021399999037384987,
"learning_rate": 4.444444444444445e-07,
"loss": 0.0005,
"step": 4800
},
{
"epoch": 9.38715953307393,
"grad_norm": 0.020681940019130707,
"learning_rate": 3.8888888888888895e-07,
"loss": 0.0004,
"step": 4825
},
{
"epoch": 9.43579766536965,
"grad_norm": 0.020026126876473427,
"learning_rate": 3.3333333333333335e-07,
"loss": 0.0005,
"step": 4850
},
{
"epoch": 9.48443579766537,
"grad_norm": 0.030826667323708534,
"learning_rate": 2.7777777777777776e-07,
"loss": 0.0004,
"step": 4875
},
{
"epoch": 9.53307392996109,
"grad_norm": 0.017975399270653725,
"learning_rate": 2.2222222222222224e-07,
"loss": 0.0004,
"step": 4900
},
{
"epoch": 9.581712062256809,
"grad_norm": 0.014251296408474445,
"learning_rate": 1.6666666666666668e-07,
"loss": 0.0005,
"step": 4925
},
{
"epoch": 9.630350194552529,
"grad_norm": 0.018128497526049614,
"learning_rate": 1.1111111111111112e-07,
"loss": 0.0005,
"step": 4950
},
{
"epoch": 9.678988326848248,
"grad_norm": 0.018901441246271133,
"learning_rate": 5.555555555555556e-08,
"loss": 0.0005,
"step": 4975
},
{
"epoch": 9.727626459143968,
"grad_norm": 0.01760680228471756,
"learning_rate": 0.0,
"loss": 0.0004,
"step": 5000
},
{
"epoch": 9.727626459143968,
"eval_loss": 0.36376532912254333,
"eval_runtime": 2157.1617,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 0.156,
"eval_wer": 0.18173684838363355,
"step": 5000
},
{
"epoch": 9.727626459143968,
"step": 5000,
"total_flos": 5.4332453168676864e+20,
"train_loss": 0.06844911024216563,
"train_runtime": 58020.7767,
"train_samples_per_second": 2.758,
"train_steps_per_second": 0.086
}
],
"logging_steps": 25,
"max_steps": 5000,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.4332453168676864e+20,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}