whisper-large-v3-ft-btb-ccv-cy / trainer_state.json
DewiBrynJones's picture
End of training
fe799c5 verified
raw
history blame
36.6 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.861003861003861,
"eval_steps": 1000,
"global_step": 5000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.019305019305019305,
"grad_norm": 7.365023136138916,
"learning_rate": 5.000000000000001e-07,
"loss": 1.4936,
"step": 25
},
{
"epoch": 0.03861003861003861,
"grad_norm": 5.869426727294922,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.2163,
"step": 50
},
{
"epoch": 0.05791505791505792,
"grad_norm": 5.041717052459717,
"learning_rate": 1.5e-06,
"loss": 0.866,
"step": 75
},
{
"epoch": 0.07722007722007722,
"grad_norm": 4.53203010559082,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.7562,
"step": 100
},
{
"epoch": 0.09652509652509653,
"grad_norm": 5.75096321105957,
"learning_rate": 2.5e-06,
"loss": 0.6816,
"step": 125
},
{
"epoch": 0.11583011583011583,
"grad_norm": 5.387002944946289,
"learning_rate": 3e-06,
"loss": 0.7012,
"step": 150
},
{
"epoch": 0.13513513513513514,
"grad_norm": 5.036501884460449,
"learning_rate": 3.5e-06,
"loss": 0.6297,
"step": 175
},
{
"epoch": 0.15444015444015444,
"grad_norm": 4.358458042144775,
"learning_rate": 4.000000000000001e-06,
"loss": 0.6081,
"step": 200
},
{
"epoch": 0.17374517374517376,
"grad_norm": 4.636687278747559,
"learning_rate": 4.5e-06,
"loss": 0.6034,
"step": 225
},
{
"epoch": 0.19305019305019305,
"grad_norm": 4.5222554206848145,
"learning_rate": 5e-06,
"loss": 0.5915,
"step": 250
},
{
"epoch": 0.21235521235521235,
"grad_norm": 4.840209007263184,
"learning_rate": 5.500000000000001e-06,
"loss": 0.6055,
"step": 275
},
{
"epoch": 0.23166023166023167,
"grad_norm": 4.3733673095703125,
"learning_rate": 6e-06,
"loss": 0.565,
"step": 300
},
{
"epoch": 0.25096525096525096,
"grad_norm": 3.8710694313049316,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.5128,
"step": 325
},
{
"epoch": 0.2702702702702703,
"grad_norm": 4.533177375793457,
"learning_rate": 7e-06,
"loss": 0.5532,
"step": 350
},
{
"epoch": 0.28957528957528955,
"grad_norm": 4.626265048980713,
"learning_rate": 7.500000000000001e-06,
"loss": 0.5708,
"step": 375
},
{
"epoch": 0.3088803088803089,
"grad_norm": 5.235887050628662,
"learning_rate": 8.000000000000001e-06,
"loss": 0.5413,
"step": 400
},
{
"epoch": 0.3281853281853282,
"grad_norm": 4.232483386993408,
"learning_rate": 8.5e-06,
"loss": 0.5326,
"step": 425
},
{
"epoch": 0.3474903474903475,
"grad_norm": 4.374448776245117,
"learning_rate": 9e-06,
"loss": 0.5205,
"step": 450
},
{
"epoch": 0.3667953667953668,
"grad_norm": 4.335531711578369,
"learning_rate": 9.5e-06,
"loss": 0.5347,
"step": 475
},
{
"epoch": 0.3861003861003861,
"grad_norm": 4.169550895690918,
"learning_rate": 1e-05,
"loss": 0.5148,
"step": 500
},
{
"epoch": 0.40540540540540543,
"grad_norm": 3.5508618354797363,
"learning_rate": 9.944444444444445e-06,
"loss": 0.5227,
"step": 525
},
{
"epoch": 0.4247104247104247,
"grad_norm": 3.7011241912841797,
"learning_rate": 9.88888888888889e-06,
"loss": 0.4989,
"step": 550
},
{
"epoch": 0.444015444015444,
"grad_norm": 3.37355899810791,
"learning_rate": 9.833333333333333e-06,
"loss": 0.5067,
"step": 575
},
{
"epoch": 0.46332046332046334,
"grad_norm": 4.071038246154785,
"learning_rate": 9.777777777777779e-06,
"loss": 0.4884,
"step": 600
},
{
"epoch": 0.4826254826254826,
"grad_norm": 3.6045796871185303,
"learning_rate": 9.722222222222223e-06,
"loss": 0.4752,
"step": 625
},
{
"epoch": 0.5019305019305019,
"grad_norm": 3.414005994796753,
"learning_rate": 9.666666666666667e-06,
"loss": 0.4784,
"step": 650
},
{
"epoch": 0.5212355212355212,
"grad_norm": 3.203928232192993,
"learning_rate": 9.611111111111112e-06,
"loss": 0.4915,
"step": 675
},
{
"epoch": 0.5405405405405406,
"grad_norm": 4.148367404937744,
"learning_rate": 9.555555555555556e-06,
"loss": 0.4861,
"step": 700
},
{
"epoch": 0.5598455598455598,
"grad_norm": 3.9569835662841797,
"learning_rate": 9.5e-06,
"loss": 0.4744,
"step": 725
},
{
"epoch": 0.5791505791505791,
"grad_norm": 3.682847261428833,
"learning_rate": 9.444444444444445e-06,
"loss": 0.4662,
"step": 750
},
{
"epoch": 0.5984555984555985,
"grad_norm": 3.498784303665161,
"learning_rate": 9.38888888888889e-06,
"loss": 0.4479,
"step": 775
},
{
"epoch": 0.6177606177606177,
"grad_norm": 3.3415608406066895,
"learning_rate": 9.333333333333334e-06,
"loss": 0.4422,
"step": 800
},
{
"epoch": 0.637065637065637,
"grad_norm": 3.7531261444091797,
"learning_rate": 9.277777777777778e-06,
"loss": 0.4639,
"step": 825
},
{
"epoch": 0.6563706563706564,
"grad_norm": 3.1370999813079834,
"learning_rate": 9.222222222222224e-06,
"loss": 0.4286,
"step": 850
},
{
"epoch": 0.6756756756756757,
"grad_norm": 3.474857807159424,
"learning_rate": 9.166666666666666e-06,
"loss": 0.4461,
"step": 875
},
{
"epoch": 0.694980694980695,
"grad_norm": 3.5445330142974854,
"learning_rate": 9.111111111111112e-06,
"loss": 0.4364,
"step": 900
},
{
"epoch": 0.7142857142857143,
"grad_norm": 3.309083938598633,
"learning_rate": 9.055555555555556e-06,
"loss": 0.4274,
"step": 925
},
{
"epoch": 0.7335907335907336,
"grad_norm": 3.169285297393799,
"learning_rate": 9e-06,
"loss": 0.4244,
"step": 950
},
{
"epoch": 0.752895752895753,
"grad_norm": 3.5987651348114014,
"learning_rate": 8.944444444444446e-06,
"loss": 0.4189,
"step": 975
},
{
"epoch": 0.7722007722007722,
"grad_norm": 3.622044563293457,
"learning_rate": 8.888888888888888e-06,
"loss": 0.4118,
"step": 1000
},
{
"epoch": 0.7722007722007722,
"eval_loss": 0.48980528116226196,
"eval_runtime": 2996.853,
"eval_samples_per_second": 2.343,
"eval_steps_per_second": 0.146,
"eval_wer": 0.3626803079679568,
"step": 1000
},
{
"epoch": 0.7915057915057915,
"grad_norm": 3.3155651092529297,
"learning_rate": 8.833333333333334e-06,
"loss": 0.4394,
"step": 1025
},
{
"epoch": 0.8108108108108109,
"grad_norm": 2.6345057487487793,
"learning_rate": 8.777777777777778e-06,
"loss": 0.4143,
"step": 1050
},
{
"epoch": 0.8301158301158301,
"grad_norm": 3.3825740814208984,
"learning_rate": 8.722222222222224e-06,
"loss": 0.399,
"step": 1075
},
{
"epoch": 0.8494208494208494,
"grad_norm": 3.439831256866455,
"learning_rate": 8.666666666666668e-06,
"loss": 0.4269,
"step": 1100
},
{
"epoch": 0.8687258687258688,
"grad_norm": 2.9175798892974854,
"learning_rate": 8.611111111111112e-06,
"loss": 0.46,
"step": 1125
},
{
"epoch": 0.888030888030888,
"grad_norm": 3.6642494201660156,
"learning_rate": 8.555555555555556e-06,
"loss": 0.4038,
"step": 1150
},
{
"epoch": 0.9073359073359073,
"grad_norm": 3.3490536212921143,
"learning_rate": 8.5e-06,
"loss": 0.3989,
"step": 1175
},
{
"epoch": 0.9266409266409267,
"grad_norm": 3.563148021697998,
"learning_rate": 8.444444444444446e-06,
"loss": 0.4221,
"step": 1200
},
{
"epoch": 0.9459459459459459,
"grad_norm": 3.2811503410339355,
"learning_rate": 8.38888888888889e-06,
"loss": 0.3959,
"step": 1225
},
{
"epoch": 0.9652509652509652,
"grad_norm": 4.630711078643799,
"learning_rate": 8.333333333333334e-06,
"loss": 0.3992,
"step": 1250
},
{
"epoch": 0.9845559845559846,
"grad_norm": 2.9398629665374756,
"learning_rate": 8.277777777777778e-06,
"loss": 0.4075,
"step": 1275
},
{
"epoch": 1.0038610038610039,
"grad_norm": 2.4827306270599365,
"learning_rate": 8.222222222222222e-06,
"loss": 0.3593,
"step": 1300
},
{
"epoch": 1.0231660231660231,
"grad_norm": 2.922198534011841,
"learning_rate": 8.166666666666668e-06,
"loss": 0.2864,
"step": 1325
},
{
"epoch": 1.0424710424710424,
"grad_norm": 2.595409393310547,
"learning_rate": 8.111111111111112e-06,
"loss": 0.2928,
"step": 1350
},
{
"epoch": 1.0617760617760619,
"grad_norm": 2.6854002475738525,
"learning_rate": 8.055555555555557e-06,
"loss": 0.2773,
"step": 1375
},
{
"epoch": 1.0810810810810811,
"grad_norm": 2.5066304206848145,
"learning_rate": 8.000000000000001e-06,
"loss": 0.2864,
"step": 1400
},
{
"epoch": 1.1003861003861004,
"grad_norm": 2.477524518966675,
"learning_rate": 7.944444444444445e-06,
"loss": 0.2815,
"step": 1425
},
{
"epoch": 1.1196911196911197,
"grad_norm": 2.947479724884033,
"learning_rate": 7.88888888888889e-06,
"loss": 0.2814,
"step": 1450
},
{
"epoch": 1.138996138996139,
"grad_norm": 2.497398853302002,
"learning_rate": 7.833333333333333e-06,
"loss": 0.27,
"step": 1475
},
{
"epoch": 1.1583011583011582,
"grad_norm": 2.5605263710021973,
"learning_rate": 7.77777777777778e-06,
"loss": 0.2778,
"step": 1500
},
{
"epoch": 1.1776061776061777,
"grad_norm": 2.8643362522125244,
"learning_rate": 7.722222222222223e-06,
"loss": 0.2625,
"step": 1525
},
{
"epoch": 1.196911196911197,
"grad_norm": 2.8837692737579346,
"learning_rate": 7.666666666666667e-06,
"loss": 0.2732,
"step": 1550
},
{
"epoch": 1.2162162162162162,
"grad_norm": 2.499385118484497,
"learning_rate": 7.611111111111111e-06,
"loss": 0.2898,
"step": 1575
},
{
"epoch": 1.2355212355212355,
"grad_norm": 3.035726547241211,
"learning_rate": 7.555555555555556e-06,
"loss": 0.279,
"step": 1600
},
{
"epoch": 1.2548262548262548,
"grad_norm": 2.872089385986328,
"learning_rate": 7.500000000000001e-06,
"loss": 0.3005,
"step": 1625
},
{
"epoch": 1.2741312741312742,
"grad_norm": 2.8907766342163086,
"learning_rate": 7.444444444444445e-06,
"loss": 0.2941,
"step": 1650
},
{
"epoch": 1.2934362934362935,
"grad_norm": 2.8525350093841553,
"learning_rate": 7.38888888888889e-06,
"loss": 0.2644,
"step": 1675
},
{
"epoch": 1.3127413127413128,
"grad_norm": 2.7391347885131836,
"learning_rate": 7.333333333333333e-06,
"loss": 0.2811,
"step": 1700
},
{
"epoch": 1.332046332046332,
"grad_norm": 2.4943618774414062,
"learning_rate": 7.277777777777778e-06,
"loss": 0.2786,
"step": 1725
},
{
"epoch": 1.3513513513513513,
"grad_norm": 3.009016752243042,
"learning_rate": 7.222222222222223e-06,
"loss": 0.2928,
"step": 1750
},
{
"epoch": 1.3706563706563706,
"grad_norm": 3.0978188514709473,
"learning_rate": 7.166666666666667e-06,
"loss": 0.298,
"step": 1775
},
{
"epoch": 1.3899613899613898,
"grad_norm": 2.746687412261963,
"learning_rate": 7.111111111111112e-06,
"loss": 0.2848,
"step": 1800
},
{
"epoch": 1.4092664092664093,
"grad_norm": 2.6545677185058594,
"learning_rate": 7.055555555555557e-06,
"loss": 0.2755,
"step": 1825
},
{
"epoch": 1.4285714285714286,
"grad_norm": 3.196070671081543,
"learning_rate": 7e-06,
"loss": 0.2733,
"step": 1850
},
{
"epoch": 1.4478764478764479,
"grad_norm": 2.5609796047210693,
"learning_rate": 6.944444444444445e-06,
"loss": 0.2659,
"step": 1875
},
{
"epoch": 1.4671814671814671,
"grad_norm": 2.339729070663452,
"learning_rate": 6.88888888888889e-06,
"loss": 0.2924,
"step": 1900
},
{
"epoch": 1.4864864864864864,
"grad_norm": 2.7920401096343994,
"learning_rate": 6.833333333333334e-06,
"loss": 0.2628,
"step": 1925
},
{
"epoch": 1.505791505791506,
"grad_norm": 2.2347285747528076,
"learning_rate": 6.777777777777779e-06,
"loss": 0.2816,
"step": 1950
},
{
"epoch": 1.525096525096525,
"grad_norm": 2.6325347423553467,
"learning_rate": 6.7222222222222235e-06,
"loss": 0.2664,
"step": 1975
},
{
"epoch": 1.5444015444015444,
"grad_norm": 3.093266248703003,
"learning_rate": 6.666666666666667e-06,
"loss": 0.2629,
"step": 2000
},
{
"epoch": 1.5444015444015444,
"eval_loss": 0.4272039532661438,
"eval_runtime": 2939.2445,
"eval_samples_per_second": 2.389,
"eval_steps_per_second": 0.149,
"eval_wer": 0.3144726762830916,
"step": 2000
},
{
"epoch": 1.5637065637065637,
"grad_norm": 2.7231810092926025,
"learning_rate": 6.6111111111111115e-06,
"loss": 0.2671,
"step": 2025
},
{
"epoch": 1.583011583011583,
"grad_norm": 2.4896557331085205,
"learning_rate": 6.555555555555556e-06,
"loss": 0.291,
"step": 2050
},
{
"epoch": 1.6023166023166024,
"grad_norm": 2.549482583999634,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.2697,
"step": 2075
},
{
"epoch": 1.6216216216216215,
"grad_norm": 3.055518627166748,
"learning_rate": 6.444444444444445e-06,
"loss": 0.2612,
"step": 2100
},
{
"epoch": 1.640926640926641,
"grad_norm": 2.8023324012756348,
"learning_rate": 6.3888888888888885e-06,
"loss": 0.2612,
"step": 2125
},
{
"epoch": 1.6602316602316602,
"grad_norm": 2.5764496326446533,
"learning_rate": 6.333333333333333e-06,
"loss": 0.2718,
"step": 2150
},
{
"epoch": 1.6795366795366795,
"grad_norm": 3.0034117698669434,
"learning_rate": 6.277777777777778e-06,
"loss": 0.2836,
"step": 2175
},
{
"epoch": 1.698841698841699,
"grad_norm": 3.285400629043579,
"learning_rate": 6.222222222222223e-06,
"loss": 0.267,
"step": 2200
},
{
"epoch": 1.718146718146718,
"grad_norm": 2.8904147148132324,
"learning_rate": 6.166666666666667e-06,
"loss": 0.2844,
"step": 2225
},
{
"epoch": 1.7374517374517375,
"grad_norm": 2.7520534992218018,
"learning_rate": 6.111111111111112e-06,
"loss": 0.2641,
"step": 2250
},
{
"epoch": 1.7567567567567568,
"grad_norm": 2.262890338897705,
"learning_rate": 6.055555555555555e-06,
"loss": 0.2636,
"step": 2275
},
{
"epoch": 1.776061776061776,
"grad_norm": 2.410085439682007,
"learning_rate": 6e-06,
"loss": 0.2753,
"step": 2300
},
{
"epoch": 1.7953667953667953,
"grad_norm": 2.5720226764678955,
"learning_rate": 5.944444444444445e-06,
"loss": 0.2702,
"step": 2325
},
{
"epoch": 1.8146718146718146,
"grad_norm": 2.999687433242798,
"learning_rate": 5.88888888888889e-06,
"loss": 0.2998,
"step": 2350
},
{
"epoch": 1.833976833976834,
"grad_norm": 2.5406405925750732,
"learning_rate": 5.833333333333334e-06,
"loss": 0.2444,
"step": 2375
},
{
"epoch": 1.8532818532818531,
"grad_norm": 2.314110517501831,
"learning_rate": 5.777777777777778e-06,
"loss": 0.2664,
"step": 2400
},
{
"epoch": 1.8725868725868726,
"grad_norm": 2.745534896850586,
"learning_rate": 5.722222222222222e-06,
"loss": 0.2676,
"step": 2425
},
{
"epoch": 1.8918918918918919,
"grad_norm": 3.1628456115722656,
"learning_rate": 5.666666666666667e-06,
"loss": 0.2475,
"step": 2450
},
{
"epoch": 1.9111969111969112,
"grad_norm": 2.6415674686431885,
"learning_rate": 5.611111111111112e-06,
"loss": 0.2565,
"step": 2475
},
{
"epoch": 1.9305019305019306,
"grad_norm": 2.4771909713745117,
"learning_rate": 5.555555555555557e-06,
"loss": 0.2621,
"step": 2500
},
{
"epoch": 1.9498069498069497,
"grad_norm": 2.7251639366149902,
"learning_rate": 5.500000000000001e-06,
"loss": 0.2703,
"step": 2525
},
{
"epoch": 1.9691119691119692,
"grad_norm": 2.73382830619812,
"learning_rate": 5.444444444444445e-06,
"loss": 0.2728,
"step": 2550
},
{
"epoch": 1.9884169884169884,
"grad_norm": 3.0017306804656982,
"learning_rate": 5.388888888888889e-06,
"loss": 0.2521,
"step": 2575
},
{
"epoch": 2.0077220077220077,
"grad_norm": 1.9712167978286743,
"learning_rate": 5.333333333333334e-06,
"loss": 0.2352,
"step": 2600
},
{
"epoch": 2.027027027027027,
"grad_norm": 1.8378448486328125,
"learning_rate": 5.2777777777777785e-06,
"loss": 0.1696,
"step": 2625
},
{
"epoch": 2.0463320463320462,
"grad_norm": 2.419578790664673,
"learning_rate": 5.2222222222222226e-06,
"loss": 0.1592,
"step": 2650
},
{
"epoch": 2.0656370656370657,
"grad_norm": 2.145256757736206,
"learning_rate": 5.1666666666666675e-06,
"loss": 0.1719,
"step": 2675
},
{
"epoch": 2.0849420849420848,
"grad_norm": 2.147104024887085,
"learning_rate": 5.1111111111111115e-06,
"loss": 0.1487,
"step": 2700
},
{
"epoch": 2.1042471042471043,
"grad_norm": 2.280949592590332,
"learning_rate": 5.0555555555555555e-06,
"loss": 0.1785,
"step": 2725
},
{
"epoch": 2.1235521235521237,
"grad_norm": 2.359393835067749,
"learning_rate": 5e-06,
"loss": 0.1656,
"step": 2750
},
{
"epoch": 2.142857142857143,
"grad_norm": 2.4849932193756104,
"learning_rate": 4.944444444444445e-06,
"loss": 0.1697,
"step": 2775
},
{
"epoch": 2.1621621621621623,
"grad_norm": 2.478525400161743,
"learning_rate": 4.888888888888889e-06,
"loss": 0.1586,
"step": 2800
},
{
"epoch": 2.1814671814671813,
"grad_norm": 2.258152723312378,
"learning_rate": 4.833333333333333e-06,
"loss": 0.1677,
"step": 2825
},
{
"epoch": 2.200772200772201,
"grad_norm": 2.2826361656188965,
"learning_rate": 4.777777777777778e-06,
"loss": 0.1603,
"step": 2850
},
{
"epoch": 2.2200772200772203,
"grad_norm": 2.3771138191223145,
"learning_rate": 4.722222222222222e-06,
"loss": 0.159,
"step": 2875
},
{
"epoch": 2.2393822393822393,
"grad_norm": 2.353187322616577,
"learning_rate": 4.666666666666667e-06,
"loss": 0.1617,
"step": 2900
},
{
"epoch": 2.258687258687259,
"grad_norm": 2.2030959129333496,
"learning_rate": 4.611111111111112e-06,
"loss": 0.17,
"step": 2925
},
{
"epoch": 2.277992277992278,
"grad_norm": 2.1587460041046143,
"learning_rate": 4.555555555555556e-06,
"loss": 0.1618,
"step": 2950
},
{
"epoch": 2.2972972972972974,
"grad_norm": 2.918156623840332,
"learning_rate": 4.5e-06,
"loss": 0.1545,
"step": 2975
},
{
"epoch": 2.3166023166023164,
"grad_norm": 2.4041364192962646,
"learning_rate": 4.444444444444444e-06,
"loss": 0.1677,
"step": 3000
},
{
"epoch": 2.3166023166023164,
"eval_loss": 0.42425960302352905,
"eval_runtime": 2930.947,
"eval_samples_per_second": 2.396,
"eval_steps_per_second": 0.15,
"eval_wer": 0.3038651796263357,
"step": 3000
},
{
"epoch": 2.335907335907336,
"grad_norm": 1.8374619483947754,
"learning_rate": 4.388888888888889e-06,
"loss": 0.1505,
"step": 3025
},
{
"epoch": 2.3552123552123554,
"grad_norm": 2.2818222045898438,
"learning_rate": 4.333333333333334e-06,
"loss": 0.1635,
"step": 3050
},
{
"epoch": 2.3745173745173744,
"grad_norm": 2.3476946353912354,
"learning_rate": 4.277777777777778e-06,
"loss": 0.1572,
"step": 3075
},
{
"epoch": 2.393822393822394,
"grad_norm": 2.727679491043091,
"learning_rate": 4.222222222222223e-06,
"loss": 0.1567,
"step": 3100
},
{
"epoch": 2.413127413127413,
"grad_norm": 2.148946523666382,
"learning_rate": 4.166666666666667e-06,
"loss": 0.1629,
"step": 3125
},
{
"epoch": 2.4324324324324325,
"grad_norm": 2.3598289489746094,
"learning_rate": 4.111111111111111e-06,
"loss": 0.1537,
"step": 3150
},
{
"epoch": 2.4517374517374515,
"grad_norm": 2.627530813217163,
"learning_rate": 4.055555555555556e-06,
"loss": 0.1617,
"step": 3175
},
{
"epoch": 2.471042471042471,
"grad_norm": 2.1012682914733887,
"learning_rate": 4.000000000000001e-06,
"loss": 0.1541,
"step": 3200
},
{
"epoch": 2.4903474903474905,
"grad_norm": 2.719054937362671,
"learning_rate": 3.944444444444445e-06,
"loss": 0.1479,
"step": 3225
},
{
"epoch": 2.5096525096525095,
"grad_norm": 1.9170578718185425,
"learning_rate": 3.88888888888889e-06,
"loss": 0.1558,
"step": 3250
},
{
"epoch": 2.528957528957529,
"grad_norm": 2.209770917892456,
"learning_rate": 3.833333333333334e-06,
"loss": 0.1494,
"step": 3275
},
{
"epoch": 2.5482625482625485,
"grad_norm": 2.3125557899475098,
"learning_rate": 3.777777777777778e-06,
"loss": 0.1549,
"step": 3300
},
{
"epoch": 2.5675675675675675,
"grad_norm": 2.2932839393615723,
"learning_rate": 3.7222222222222225e-06,
"loss": 0.1529,
"step": 3325
},
{
"epoch": 2.586872586872587,
"grad_norm": 2.347710371017456,
"learning_rate": 3.6666666666666666e-06,
"loss": 0.1637,
"step": 3350
},
{
"epoch": 2.606177606177606,
"grad_norm": 2.479405641555786,
"learning_rate": 3.6111111111111115e-06,
"loss": 0.162,
"step": 3375
},
{
"epoch": 2.6254826254826256,
"grad_norm": 2.7078042030334473,
"learning_rate": 3.555555555555556e-06,
"loss": 0.1606,
"step": 3400
},
{
"epoch": 2.6447876447876446,
"grad_norm": 2.097592830657959,
"learning_rate": 3.5e-06,
"loss": 0.1554,
"step": 3425
},
{
"epoch": 2.664092664092664,
"grad_norm": 2.405099630355835,
"learning_rate": 3.444444444444445e-06,
"loss": 0.1611,
"step": 3450
},
{
"epoch": 2.6833976833976836,
"grad_norm": 2.120976686477661,
"learning_rate": 3.3888888888888893e-06,
"loss": 0.1537,
"step": 3475
},
{
"epoch": 2.7027027027027026,
"grad_norm": 2.165222406387329,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.156,
"step": 3500
},
{
"epoch": 2.722007722007722,
"grad_norm": 2.3541574478149414,
"learning_rate": 3.277777777777778e-06,
"loss": 0.1571,
"step": 3525
},
{
"epoch": 2.741312741312741,
"grad_norm": 2.5069708824157715,
"learning_rate": 3.2222222222222227e-06,
"loss": 0.1467,
"step": 3550
},
{
"epoch": 2.7606177606177607,
"grad_norm": 1.9154902696609497,
"learning_rate": 3.1666666666666667e-06,
"loss": 0.1556,
"step": 3575
},
{
"epoch": 2.7799227799227797,
"grad_norm": 1.964211106300354,
"learning_rate": 3.1111111111111116e-06,
"loss": 0.1534,
"step": 3600
},
{
"epoch": 2.799227799227799,
"grad_norm": 2.3793370723724365,
"learning_rate": 3.055555555555556e-06,
"loss": 0.1578,
"step": 3625
},
{
"epoch": 2.8185328185328187,
"grad_norm": 2.7694272994995117,
"learning_rate": 3e-06,
"loss": 0.1518,
"step": 3650
},
{
"epoch": 2.8378378378378377,
"grad_norm": 2.1063687801361084,
"learning_rate": 2.944444444444445e-06,
"loss": 0.1534,
"step": 3675
},
{
"epoch": 2.857142857142857,
"grad_norm": 2.708767890930176,
"learning_rate": 2.888888888888889e-06,
"loss": 0.153,
"step": 3700
},
{
"epoch": 2.8764478764478767,
"grad_norm": 1.7426552772521973,
"learning_rate": 2.8333333333333335e-06,
"loss": 0.1501,
"step": 3725
},
{
"epoch": 2.8957528957528957,
"grad_norm": 2.721165180206299,
"learning_rate": 2.7777777777777783e-06,
"loss": 0.1613,
"step": 3750
},
{
"epoch": 2.915057915057915,
"grad_norm": 2.525545358657837,
"learning_rate": 2.7222222222222224e-06,
"loss": 0.1534,
"step": 3775
},
{
"epoch": 2.9343629343629343,
"grad_norm": 2.6334362030029297,
"learning_rate": 2.666666666666667e-06,
"loss": 0.158,
"step": 3800
},
{
"epoch": 2.9536679536679538,
"grad_norm": 2.1491916179656982,
"learning_rate": 2.6111111111111113e-06,
"loss": 0.15,
"step": 3825
},
{
"epoch": 2.972972972972973,
"grad_norm": 2.0909786224365234,
"learning_rate": 2.5555555555555557e-06,
"loss": 0.147,
"step": 3850
},
{
"epoch": 2.9922779922779923,
"grad_norm": 2.8027398586273193,
"learning_rate": 2.5e-06,
"loss": 0.1642,
"step": 3875
},
{
"epoch": 3.011583011583012,
"grad_norm": 1.8371341228485107,
"learning_rate": 2.4444444444444447e-06,
"loss": 0.1176,
"step": 3900
},
{
"epoch": 3.030888030888031,
"grad_norm": 2.1273913383483887,
"learning_rate": 2.388888888888889e-06,
"loss": 0.0937,
"step": 3925
},
{
"epoch": 3.0501930501930503,
"grad_norm": 1.8209782838821411,
"learning_rate": 2.3333333333333336e-06,
"loss": 0.089,
"step": 3950
},
{
"epoch": 3.0694980694980694,
"grad_norm": 2.118212938308716,
"learning_rate": 2.277777777777778e-06,
"loss": 0.0888,
"step": 3975
},
{
"epoch": 3.088803088803089,
"grad_norm": 1.675663709640503,
"learning_rate": 2.222222222222222e-06,
"loss": 0.0841,
"step": 4000
},
{
"epoch": 3.088803088803089,
"eval_loss": 0.4492855668067932,
"eval_runtime": 2954.4178,
"eval_samples_per_second": 2.377,
"eval_steps_per_second": 0.149,
"eval_wer": 0.2968411212526454,
"step": 4000
},
{
"epoch": 3.108108108108108,
"grad_norm": 2.0060861110687256,
"learning_rate": 2.166666666666667e-06,
"loss": 0.084,
"step": 4025
},
{
"epoch": 3.1274131274131274,
"grad_norm": 1.8335351943969727,
"learning_rate": 2.1111111111111114e-06,
"loss": 0.0871,
"step": 4050
},
{
"epoch": 3.146718146718147,
"grad_norm": 2.165919303894043,
"learning_rate": 2.0555555555555555e-06,
"loss": 0.0892,
"step": 4075
},
{
"epoch": 3.166023166023166,
"grad_norm": 1.5656211376190186,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.0869,
"step": 4100
},
{
"epoch": 3.1853281853281854,
"grad_norm": 1.9687210321426392,
"learning_rate": 1.944444444444445e-06,
"loss": 0.0878,
"step": 4125
},
{
"epoch": 3.2046332046332044,
"grad_norm": 2.017395496368408,
"learning_rate": 1.888888888888889e-06,
"loss": 0.0823,
"step": 4150
},
{
"epoch": 3.223938223938224,
"grad_norm": 2.3069615364074707,
"learning_rate": 1.8333333333333333e-06,
"loss": 0.0894,
"step": 4175
},
{
"epoch": 3.2432432432432434,
"grad_norm": 1.8950186967849731,
"learning_rate": 1.777777777777778e-06,
"loss": 0.0889,
"step": 4200
},
{
"epoch": 3.2625482625482625,
"grad_norm": 1.9198052883148193,
"learning_rate": 1.7222222222222224e-06,
"loss": 0.081,
"step": 4225
},
{
"epoch": 3.281853281853282,
"grad_norm": 2.25732421875,
"learning_rate": 1.6666666666666667e-06,
"loss": 0.0872,
"step": 4250
},
{
"epoch": 3.301158301158301,
"grad_norm": 1.6695570945739746,
"learning_rate": 1.6111111111111113e-06,
"loss": 0.0827,
"step": 4275
},
{
"epoch": 3.3204633204633205,
"grad_norm": 2.1860039234161377,
"learning_rate": 1.5555555555555558e-06,
"loss": 0.0864,
"step": 4300
},
{
"epoch": 3.33976833976834,
"grad_norm": 2.0618574619293213,
"learning_rate": 1.5e-06,
"loss": 0.0843,
"step": 4325
},
{
"epoch": 3.359073359073359,
"grad_norm": 1.952563762664795,
"learning_rate": 1.4444444444444445e-06,
"loss": 0.0822,
"step": 4350
},
{
"epoch": 3.3783783783783785,
"grad_norm": 2.2584211826324463,
"learning_rate": 1.3888888888888892e-06,
"loss": 0.087,
"step": 4375
},
{
"epoch": 3.3976833976833976,
"grad_norm": 1.7842656373977661,
"learning_rate": 1.3333333333333334e-06,
"loss": 0.0856,
"step": 4400
},
{
"epoch": 3.416988416988417,
"grad_norm": 2.130709409713745,
"learning_rate": 1.2777777777777779e-06,
"loss": 0.0886,
"step": 4425
},
{
"epoch": 3.436293436293436,
"grad_norm": 1.505071759223938,
"learning_rate": 1.2222222222222223e-06,
"loss": 0.0795,
"step": 4450
},
{
"epoch": 3.4555984555984556,
"grad_norm": 2.2736964225769043,
"learning_rate": 1.1666666666666668e-06,
"loss": 0.0802,
"step": 4475
},
{
"epoch": 3.474903474903475,
"grad_norm": 1.685120940208435,
"learning_rate": 1.111111111111111e-06,
"loss": 0.085,
"step": 4500
},
{
"epoch": 3.494208494208494,
"grad_norm": 1.8761180639266968,
"learning_rate": 1.0555555555555557e-06,
"loss": 0.0732,
"step": 4525
},
{
"epoch": 3.5135135135135136,
"grad_norm": 1.7968302965164185,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.0868,
"step": 4550
},
{
"epoch": 3.532818532818533,
"grad_norm": 1.6335569620132446,
"learning_rate": 9.444444444444445e-07,
"loss": 0.0787,
"step": 4575
},
{
"epoch": 3.552123552123552,
"grad_norm": 1.402917742729187,
"learning_rate": 8.88888888888889e-07,
"loss": 0.0779,
"step": 4600
},
{
"epoch": 3.571428571428571,
"grad_norm": 1.4878441095352173,
"learning_rate": 8.333333333333333e-07,
"loss": 0.0795,
"step": 4625
},
{
"epoch": 3.5907335907335907,
"grad_norm": 1.846757411956787,
"learning_rate": 7.777777777777779e-07,
"loss": 0.0868,
"step": 4650
},
{
"epoch": 3.61003861003861,
"grad_norm": 1.496010422706604,
"learning_rate": 7.222222222222222e-07,
"loss": 0.0836,
"step": 4675
},
{
"epoch": 3.629343629343629,
"grad_norm": 2.0897486209869385,
"learning_rate": 6.666666666666667e-07,
"loss": 0.0839,
"step": 4700
},
{
"epoch": 3.6486486486486487,
"grad_norm": 1.6161247491836548,
"learning_rate": 6.111111111111112e-07,
"loss": 0.0858,
"step": 4725
},
{
"epoch": 3.667953667953668,
"grad_norm": 1.9887114763259888,
"learning_rate": 5.555555555555555e-07,
"loss": 0.0864,
"step": 4750
},
{
"epoch": 3.687258687258687,
"grad_norm": 2.271867513656616,
"learning_rate": 5.000000000000001e-07,
"loss": 0.0815,
"step": 4775
},
{
"epoch": 3.7065637065637067,
"grad_norm": 1.9788146018981934,
"learning_rate": 4.444444444444445e-07,
"loss": 0.0821,
"step": 4800
},
{
"epoch": 3.7258687258687258,
"grad_norm": 1.8956130743026733,
"learning_rate": 3.8888888888888895e-07,
"loss": 0.0789,
"step": 4825
},
{
"epoch": 3.7451737451737452,
"grad_norm": 2.4207746982574463,
"learning_rate": 3.3333333333333335e-07,
"loss": 0.1009,
"step": 4850
},
{
"epoch": 3.7644787644787643,
"grad_norm": 1.7667814493179321,
"learning_rate": 2.7777777777777776e-07,
"loss": 0.0836,
"step": 4875
},
{
"epoch": 3.7837837837837838,
"grad_norm": 2.118232488632202,
"learning_rate": 2.2222222222222224e-07,
"loss": 0.0824,
"step": 4900
},
{
"epoch": 3.8030888030888033,
"grad_norm": 1.5285857915878296,
"learning_rate": 1.6666666666666668e-07,
"loss": 0.0836,
"step": 4925
},
{
"epoch": 3.8223938223938223,
"grad_norm": 2.041189670562744,
"learning_rate": 1.1111111111111112e-07,
"loss": 0.0838,
"step": 4950
},
{
"epoch": 3.841698841698842,
"grad_norm": 2.062230110168457,
"learning_rate": 5.555555555555556e-08,
"loss": 0.0804,
"step": 4975
},
{
"epoch": 3.861003861003861,
"grad_norm": 1.9036461114883423,
"learning_rate": 0.0,
"loss": 0.0845,
"step": 5000
},
{
"epoch": 3.861003861003861,
"eval_loss": 0.447899729013443,
"eval_runtime": 2978.0577,
"eval_samples_per_second": 2.358,
"eval_steps_per_second": 0.147,
"eval_wer": 0.29472481531011024,
"step": 5000
},
{
"epoch": 3.861003861003861,
"step": 5000,
"total_flos": 5.435487665750016e+20,
"train_loss": 0.2695653033494949,
"train_runtime": 64824.1008,
"train_samples_per_second": 2.468,
"train_steps_per_second": 0.077
}
],
"logging_steps": 25,
"max_steps": 5000,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.435487665750016e+20,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}