whisper-large-v3-ft-btb-cy / trainer_state.json
DewiBrynJones's picture
End of training
91c0069 verified
raw
history blame
36.5 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.694760820045558,
"eval_steps": 1000,
"global_step": 5000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02847380410022779,
"grad_norm": 8.5387601852417,
"learning_rate": 5.000000000000001e-07,
"loss": 1.7657,
"step": 25
},
{
"epoch": 0.05694760820045558,
"grad_norm": 5.922762393951416,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.3302,
"step": 50
},
{
"epoch": 0.08542141230068337,
"grad_norm": 6.087216854095459,
"learning_rate": 1.5e-06,
"loss": 0.9581,
"step": 75
},
{
"epoch": 0.11389521640091116,
"grad_norm": 5.635214328765869,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.8187,
"step": 100
},
{
"epoch": 0.14236902050113895,
"grad_norm": 4.90886926651001,
"learning_rate": 2.5e-06,
"loss": 0.7346,
"step": 125
},
{
"epoch": 0.17084282460136674,
"grad_norm": 5.0317230224609375,
"learning_rate": 3e-06,
"loss": 0.7124,
"step": 150
},
{
"epoch": 0.19931662870159453,
"grad_norm": 5.094836711883545,
"learning_rate": 3.5e-06,
"loss": 0.6898,
"step": 175
},
{
"epoch": 0.22779043280182232,
"grad_norm": 4.328588962554932,
"learning_rate": 4.000000000000001e-06,
"loss": 0.649,
"step": 200
},
{
"epoch": 0.25626423690205014,
"grad_norm": 5.198021411895752,
"learning_rate": 4.5e-06,
"loss": 0.6179,
"step": 225
},
{
"epoch": 0.2847380410022779,
"grad_norm": 4.642407417297363,
"learning_rate": 5e-06,
"loss": 0.6448,
"step": 250
},
{
"epoch": 0.3132118451025057,
"grad_norm": 4.232536315917969,
"learning_rate": 5.500000000000001e-06,
"loss": 0.6087,
"step": 275
},
{
"epoch": 0.3416856492027335,
"grad_norm": 4.383147239685059,
"learning_rate": 6e-06,
"loss": 0.5977,
"step": 300
},
{
"epoch": 0.3701594533029613,
"grad_norm": 4.431395530700684,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.6024,
"step": 325
},
{
"epoch": 0.39863325740318906,
"grad_norm": 5.041561126708984,
"learning_rate": 7e-06,
"loss": 0.5695,
"step": 350
},
{
"epoch": 0.4271070615034169,
"grad_norm": 4.403401851654053,
"learning_rate": 7.500000000000001e-06,
"loss": 0.5453,
"step": 375
},
{
"epoch": 0.45558086560364464,
"grad_norm": 4.56962251663208,
"learning_rate": 8.000000000000001e-06,
"loss": 0.5746,
"step": 400
},
{
"epoch": 0.48405466970387245,
"grad_norm": 4.166983127593994,
"learning_rate": 8.5e-06,
"loss": 0.5548,
"step": 425
},
{
"epoch": 0.5125284738041003,
"grad_norm": 3.995704174041748,
"learning_rate": 9e-06,
"loss": 0.5251,
"step": 450
},
{
"epoch": 0.541002277904328,
"grad_norm": 4.352871417999268,
"learning_rate": 9.5e-06,
"loss": 0.529,
"step": 475
},
{
"epoch": 0.5694760820045558,
"grad_norm": 5.0144219398498535,
"learning_rate": 1e-05,
"loss": 0.5356,
"step": 500
},
{
"epoch": 0.5979498861047836,
"grad_norm": 4.578596591949463,
"learning_rate": 9.944444444444445e-06,
"loss": 0.5513,
"step": 525
},
{
"epoch": 0.6264236902050114,
"grad_norm": 4.121158123016357,
"learning_rate": 9.88888888888889e-06,
"loss": 0.5373,
"step": 550
},
{
"epoch": 0.6548974943052391,
"grad_norm": 3.4131343364715576,
"learning_rate": 9.833333333333333e-06,
"loss": 0.5357,
"step": 575
},
{
"epoch": 0.683371298405467,
"grad_norm": 3.698122501373291,
"learning_rate": 9.777777777777779e-06,
"loss": 0.5219,
"step": 600
},
{
"epoch": 0.7118451025056948,
"grad_norm": 3.6946487426757812,
"learning_rate": 9.722222222222223e-06,
"loss": 0.525,
"step": 625
},
{
"epoch": 0.7403189066059226,
"grad_norm": 3.6262104511260986,
"learning_rate": 9.666666666666667e-06,
"loss": 0.5085,
"step": 650
},
{
"epoch": 0.7687927107061503,
"grad_norm": 2.9753031730651855,
"learning_rate": 9.611111111111112e-06,
"loss": 0.4876,
"step": 675
},
{
"epoch": 0.7972665148063781,
"grad_norm": 3.8146708011627197,
"learning_rate": 9.555555555555556e-06,
"loss": 0.5017,
"step": 700
},
{
"epoch": 0.8257403189066059,
"grad_norm": 3.672109603881836,
"learning_rate": 9.5e-06,
"loss": 0.4733,
"step": 725
},
{
"epoch": 0.8542141230068337,
"grad_norm": 4.09952974319458,
"learning_rate": 9.444444444444445e-06,
"loss": 0.4667,
"step": 750
},
{
"epoch": 0.8826879271070615,
"grad_norm": 3.828016519546509,
"learning_rate": 9.38888888888889e-06,
"loss": 0.468,
"step": 775
},
{
"epoch": 0.9111617312072893,
"grad_norm": 3.191138744354248,
"learning_rate": 9.333333333333334e-06,
"loss": 0.4772,
"step": 800
},
{
"epoch": 0.9396355353075171,
"grad_norm": 4.035594463348389,
"learning_rate": 9.277777777777778e-06,
"loss": 0.4515,
"step": 825
},
{
"epoch": 0.9681093394077449,
"grad_norm": 3.4418606758117676,
"learning_rate": 9.222222222222224e-06,
"loss": 0.4821,
"step": 850
},
{
"epoch": 0.9965831435079726,
"grad_norm": 3.5060410499572754,
"learning_rate": 9.166666666666666e-06,
"loss": 0.4541,
"step": 875
},
{
"epoch": 1.0250569476082005,
"grad_norm": 2.4985129833221436,
"learning_rate": 9.111111111111112e-06,
"loss": 0.3697,
"step": 900
},
{
"epoch": 1.0535307517084282,
"grad_norm": 3.486703395843506,
"learning_rate": 9.055555555555556e-06,
"loss": 0.3357,
"step": 925
},
{
"epoch": 1.082004555808656,
"grad_norm": 2.972067356109619,
"learning_rate": 9e-06,
"loss": 0.344,
"step": 950
},
{
"epoch": 1.1104783599088839,
"grad_norm": 3.0161571502685547,
"learning_rate": 8.944444444444446e-06,
"loss": 0.3496,
"step": 975
},
{
"epoch": 1.1389521640091116,
"grad_norm": 3.0906615257263184,
"learning_rate": 8.888888888888888e-06,
"loss": 0.3452,
"step": 1000
},
{
"epoch": 1.1389521640091116,
"eval_loss": 0.4483538568019867,
"eval_runtime": 2994.5795,
"eval_samples_per_second": 2.345,
"eval_steps_per_second": 0.147,
"eval_wer": 0.31664895030560725,
"step": 1000
},
{
"epoch": 1.1674259681093395,
"grad_norm": 3.4060277938842773,
"learning_rate": 8.833333333333334e-06,
"loss": 0.3284,
"step": 1025
},
{
"epoch": 1.1958997722095672,
"grad_norm": 3.258347988128662,
"learning_rate": 8.777777777777778e-06,
"loss": 0.342,
"step": 1050
},
{
"epoch": 1.224373576309795,
"grad_norm": 3.067859172821045,
"learning_rate": 8.722222222222224e-06,
"loss": 0.3313,
"step": 1075
},
{
"epoch": 1.2528473804100229,
"grad_norm": 3.383387804031372,
"learning_rate": 8.666666666666668e-06,
"loss": 0.3345,
"step": 1100
},
{
"epoch": 1.2813211845102506,
"grad_norm": 2.9198620319366455,
"learning_rate": 8.611111111111112e-06,
"loss": 0.3493,
"step": 1125
},
{
"epoch": 1.3097949886104785,
"grad_norm": 2.705611228942871,
"learning_rate": 8.555555555555556e-06,
"loss": 0.3438,
"step": 1150
},
{
"epoch": 1.3382687927107062,
"grad_norm": 3.0571541786193848,
"learning_rate": 8.5e-06,
"loss": 0.3258,
"step": 1175
},
{
"epoch": 1.366742596810934,
"grad_norm": 3.117966890335083,
"learning_rate": 8.444444444444446e-06,
"loss": 0.3437,
"step": 1200
},
{
"epoch": 1.3952164009111616,
"grad_norm": 2.608323812484741,
"learning_rate": 8.38888888888889e-06,
"loss": 0.3287,
"step": 1225
},
{
"epoch": 1.4236902050113895,
"grad_norm": 2.7301223278045654,
"learning_rate": 8.333333333333334e-06,
"loss": 0.3319,
"step": 1250
},
{
"epoch": 1.4521640091116172,
"grad_norm": 2.6983442306518555,
"learning_rate": 8.277777777777778e-06,
"loss": 0.3139,
"step": 1275
},
{
"epoch": 1.4806378132118452,
"grad_norm": 3.1055397987365723,
"learning_rate": 8.222222222222222e-06,
"loss": 0.3596,
"step": 1300
},
{
"epoch": 1.5091116173120729,
"grad_norm": 2.650721311569214,
"learning_rate": 8.166666666666668e-06,
"loss": 0.3387,
"step": 1325
},
{
"epoch": 1.5375854214123006,
"grad_norm": 2.6098825931549072,
"learning_rate": 8.111111111111112e-06,
"loss": 0.3329,
"step": 1350
},
{
"epoch": 1.5660592255125285,
"grad_norm": 2.5943613052368164,
"learning_rate": 8.055555555555557e-06,
"loss": 0.34,
"step": 1375
},
{
"epoch": 1.5945330296127562,
"grad_norm": 3.092444896697998,
"learning_rate": 8.000000000000001e-06,
"loss": 0.3329,
"step": 1400
},
{
"epoch": 1.6230068337129842,
"grad_norm": 2.703401565551758,
"learning_rate": 7.944444444444445e-06,
"loss": 0.3309,
"step": 1425
},
{
"epoch": 1.6514806378132119,
"grad_norm": 3.2039897441864014,
"learning_rate": 7.88888888888889e-06,
"loss": 0.3204,
"step": 1450
},
{
"epoch": 1.6799544419134396,
"grad_norm": 2.8489060401916504,
"learning_rate": 7.833333333333333e-06,
"loss": 0.341,
"step": 1475
},
{
"epoch": 1.7084282460136673,
"grad_norm": 3.1537883281707764,
"learning_rate": 7.77777777777778e-06,
"loss": 0.3361,
"step": 1500
},
{
"epoch": 1.7369020501138952,
"grad_norm": 3.1526262760162354,
"learning_rate": 7.722222222222223e-06,
"loss": 0.3253,
"step": 1525
},
{
"epoch": 1.7653758542141231,
"grad_norm": 3.4786808490753174,
"learning_rate": 7.666666666666667e-06,
"loss": 0.3271,
"step": 1550
},
{
"epoch": 1.7938496583143508,
"grad_norm": 2.7717833518981934,
"learning_rate": 7.611111111111111e-06,
"loss": 0.3328,
"step": 1575
},
{
"epoch": 1.8223234624145785,
"grad_norm": 2.6981112957000732,
"learning_rate": 7.555555555555556e-06,
"loss": 0.3314,
"step": 1600
},
{
"epoch": 1.8507972665148062,
"grad_norm": 3.0883076190948486,
"learning_rate": 7.500000000000001e-06,
"loss": 0.35,
"step": 1625
},
{
"epoch": 1.8792710706150342,
"grad_norm": 3.0751681327819824,
"learning_rate": 7.444444444444445e-06,
"loss": 0.3256,
"step": 1650
},
{
"epoch": 1.907744874715262,
"grad_norm": 2.925309658050537,
"learning_rate": 7.38888888888889e-06,
"loss": 0.3386,
"step": 1675
},
{
"epoch": 1.9362186788154898,
"grad_norm": 3.185584545135498,
"learning_rate": 7.333333333333333e-06,
"loss": 0.315,
"step": 1700
},
{
"epoch": 1.9646924829157175,
"grad_norm": 3.3560593128204346,
"learning_rate": 7.277777777777778e-06,
"loss": 0.3273,
"step": 1725
},
{
"epoch": 1.9931662870159452,
"grad_norm": 2.6927008628845215,
"learning_rate": 7.222222222222223e-06,
"loss": 0.3162,
"step": 1750
},
{
"epoch": 2.021640091116173,
"grad_norm": 2.982485771179199,
"learning_rate": 7.166666666666667e-06,
"loss": 0.2377,
"step": 1775
},
{
"epoch": 2.050113895216401,
"grad_norm": 2.3905413150787354,
"learning_rate": 7.111111111111112e-06,
"loss": 0.2009,
"step": 1800
},
{
"epoch": 2.078587699316629,
"grad_norm": 2.3832273483276367,
"learning_rate": 7.055555555555557e-06,
"loss": 0.1986,
"step": 1825
},
{
"epoch": 2.1070615034168565,
"grad_norm": 2.6094419956207275,
"learning_rate": 7e-06,
"loss": 0.2018,
"step": 1850
},
{
"epoch": 2.135535307517084,
"grad_norm": 2.6407711505889893,
"learning_rate": 6.944444444444445e-06,
"loss": 0.1946,
"step": 1875
},
{
"epoch": 2.164009111617312,
"grad_norm": 2.179900884628296,
"learning_rate": 6.88888888888889e-06,
"loss": 0.197,
"step": 1900
},
{
"epoch": 2.19248291571754,
"grad_norm": 2.5007543563842773,
"learning_rate": 6.833333333333334e-06,
"loss": 0.2,
"step": 1925
},
{
"epoch": 2.2209567198177678,
"grad_norm": 2.886491537094116,
"learning_rate": 6.777777777777779e-06,
"loss": 0.2098,
"step": 1950
},
{
"epoch": 2.2494305239179955,
"grad_norm": 3.0158777236938477,
"learning_rate": 6.7222222222222235e-06,
"loss": 0.2073,
"step": 1975
},
{
"epoch": 2.277904328018223,
"grad_norm": 2.4290666580200195,
"learning_rate": 6.666666666666667e-06,
"loss": 0.2029,
"step": 2000
},
{
"epoch": 2.277904328018223,
"eval_loss": 0.4125473201274872,
"eval_runtime": 2968.6025,
"eval_samples_per_second": 2.365,
"eval_steps_per_second": 0.148,
"eval_wer": 0.28979537602976346,
"step": 2000
},
{
"epoch": 2.306378132118451,
"grad_norm": 2.810905694961548,
"learning_rate": 6.6111111111111115e-06,
"loss": 0.1981,
"step": 2025
},
{
"epoch": 2.334851936218679,
"grad_norm": 2.6154942512512207,
"learning_rate": 6.555555555555556e-06,
"loss": 0.2148,
"step": 2050
},
{
"epoch": 2.3633257403189067,
"grad_norm": 2.636270761489868,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.1926,
"step": 2075
},
{
"epoch": 2.3917995444191344,
"grad_norm": 2.7535417079925537,
"learning_rate": 6.444444444444445e-06,
"loss": 0.2098,
"step": 2100
},
{
"epoch": 2.420273348519362,
"grad_norm": 2.86851167678833,
"learning_rate": 6.3888888888888885e-06,
"loss": 0.2182,
"step": 2125
},
{
"epoch": 2.44874715261959,
"grad_norm": 2.129084348678589,
"learning_rate": 6.333333333333333e-06,
"loss": 0.2137,
"step": 2150
},
{
"epoch": 2.477220956719818,
"grad_norm": 2.456988573074341,
"learning_rate": 6.277777777777778e-06,
"loss": 0.1998,
"step": 2175
},
{
"epoch": 2.5056947608200457,
"grad_norm": 2.829132080078125,
"learning_rate": 6.222222222222223e-06,
"loss": 0.2076,
"step": 2200
},
{
"epoch": 2.5341685649202734,
"grad_norm": 3.0912461280822754,
"learning_rate": 6.166666666666667e-06,
"loss": 0.2047,
"step": 2225
},
{
"epoch": 2.562642369020501,
"grad_norm": 2.693300247192383,
"learning_rate": 6.111111111111112e-06,
"loss": 0.2043,
"step": 2250
},
{
"epoch": 2.591116173120729,
"grad_norm": 2.593743085861206,
"learning_rate": 6.055555555555555e-06,
"loss": 0.2023,
"step": 2275
},
{
"epoch": 2.619589977220957,
"grad_norm": 3.2144663333892822,
"learning_rate": 6e-06,
"loss": 0.2013,
"step": 2300
},
{
"epoch": 2.6480637813211843,
"grad_norm": 2.123487710952759,
"learning_rate": 5.944444444444445e-06,
"loss": 0.2129,
"step": 2325
},
{
"epoch": 2.6765375854214124,
"grad_norm": 2.5355987548828125,
"learning_rate": 5.88888888888889e-06,
"loss": 0.1932,
"step": 2350
},
{
"epoch": 2.70501138952164,
"grad_norm": 2.0367653369903564,
"learning_rate": 5.833333333333334e-06,
"loss": 0.2015,
"step": 2375
},
{
"epoch": 2.733485193621868,
"grad_norm": 2.442157506942749,
"learning_rate": 5.777777777777778e-06,
"loss": 0.1905,
"step": 2400
},
{
"epoch": 2.7619589977220955,
"grad_norm": 2.7985150814056396,
"learning_rate": 5.722222222222222e-06,
"loss": 0.2093,
"step": 2425
},
{
"epoch": 2.7904328018223232,
"grad_norm": 2.2095937728881836,
"learning_rate": 5.666666666666667e-06,
"loss": 0.196,
"step": 2450
},
{
"epoch": 2.8189066059225514,
"grad_norm": 2.453608751296997,
"learning_rate": 5.611111111111112e-06,
"loss": 0.2113,
"step": 2475
},
{
"epoch": 2.847380410022779,
"grad_norm": 2.1963589191436768,
"learning_rate": 5.555555555555557e-06,
"loss": 0.1997,
"step": 2500
},
{
"epoch": 2.875854214123007,
"grad_norm": 2.492926836013794,
"learning_rate": 5.500000000000001e-06,
"loss": 0.2123,
"step": 2525
},
{
"epoch": 2.9043280182232345,
"grad_norm": 2.3640942573547363,
"learning_rate": 5.444444444444445e-06,
"loss": 0.1959,
"step": 2550
},
{
"epoch": 2.932801822323462,
"grad_norm": 2.2937705516815186,
"learning_rate": 5.388888888888889e-06,
"loss": 0.1879,
"step": 2575
},
{
"epoch": 2.9612756264236904,
"grad_norm": 2.901215076446533,
"learning_rate": 5.333333333333334e-06,
"loss": 0.2008,
"step": 2600
},
{
"epoch": 2.989749430523918,
"grad_norm": 2.2094128131866455,
"learning_rate": 5.2777777777777785e-06,
"loss": 0.2039,
"step": 2625
},
{
"epoch": 3.0182232346241458,
"grad_norm": 1.8943818807601929,
"learning_rate": 5.2222222222222226e-06,
"loss": 0.1459,
"step": 2650
},
{
"epoch": 3.0466970387243735,
"grad_norm": 2.121443033218384,
"learning_rate": 5.1666666666666675e-06,
"loss": 0.1153,
"step": 2675
},
{
"epoch": 3.075170842824601,
"grad_norm": 2.070572853088379,
"learning_rate": 5.1111111111111115e-06,
"loss": 0.1065,
"step": 2700
},
{
"epoch": 3.1036446469248293,
"grad_norm": 1.9575121402740479,
"learning_rate": 5.0555555555555555e-06,
"loss": 0.1104,
"step": 2725
},
{
"epoch": 3.132118451025057,
"grad_norm": 2.5714454650878906,
"learning_rate": 5e-06,
"loss": 0.1152,
"step": 2750
},
{
"epoch": 3.1605922551252847,
"grad_norm": 1.941654920578003,
"learning_rate": 4.944444444444445e-06,
"loss": 0.1156,
"step": 2775
},
{
"epoch": 3.1890660592255125,
"grad_norm": 2.314999580383301,
"learning_rate": 4.888888888888889e-06,
"loss": 0.1079,
"step": 2800
},
{
"epoch": 3.21753986332574,
"grad_norm": 2.4355452060699463,
"learning_rate": 4.833333333333333e-06,
"loss": 0.1267,
"step": 2825
},
{
"epoch": 3.2460136674259683,
"grad_norm": 2.246598720550537,
"learning_rate": 4.777777777777778e-06,
"loss": 0.1036,
"step": 2850
},
{
"epoch": 3.274487471526196,
"grad_norm": 1.9095807075500488,
"learning_rate": 4.722222222222222e-06,
"loss": 0.1124,
"step": 2875
},
{
"epoch": 3.3029612756264237,
"grad_norm": 2.485973596572876,
"learning_rate": 4.666666666666667e-06,
"loss": 0.115,
"step": 2900
},
{
"epoch": 3.3314350797266514,
"grad_norm": 2.0243821144104004,
"learning_rate": 4.611111111111112e-06,
"loss": 0.1073,
"step": 2925
},
{
"epoch": 3.359908883826879,
"grad_norm": 2.4946811199188232,
"learning_rate": 4.555555555555556e-06,
"loss": 0.1127,
"step": 2950
},
{
"epoch": 3.3883826879271073,
"grad_norm": 2.0500540733337402,
"learning_rate": 4.5e-06,
"loss": 0.1085,
"step": 2975
},
{
"epoch": 3.416856492027335,
"grad_norm": 2.5584237575531006,
"learning_rate": 4.444444444444444e-06,
"loss": 0.1143,
"step": 3000
},
{
"epoch": 3.416856492027335,
"eval_loss": 0.432865709066391,
"eval_runtime": 2970.1299,
"eval_samples_per_second": 2.364,
"eval_steps_per_second": 0.148,
"eval_wer": 0.28141110815838427,
"step": 3000
},
{
"epoch": 3.4453302961275627,
"grad_norm": 2.3055500984191895,
"learning_rate": 4.388888888888889e-06,
"loss": 0.1135,
"step": 3025
},
{
"epoch": 3.4738041002277904,
"grad_norm": 2.1304893493652344,
"learning_rate": 4.333333333333334e-06,
"loss": 0.1042,
"step": 3050
},
{
"epoch": 3.502277904328018,
"grad_norm": 2.5304181575775146,
"learning_rate": 4.277777777777778e-06,
"loss": 0.1156,
"step": 3075
},
{
"epoch": 3.5307517084282463,
"grad_norm": 2.2031352519989014,
"learning_rate": 4.222222222222223e-06,
"loss": 0.1104,
"step": 3100
},
{
"epoch": 3.559225512528474,
"grad_norm": 2.125488758087158,
"learning_rate": 4.166666666666667e-06,
"loss": 0.1114,
"step": 3125
},
{
"epoch": 3.5876993166287017,
"grad_norm": 2.4652950763702393,
"learning_rate": 4.111111111111111e-06,
"loss": 0.1094,
"step": 3150
},
{
"epoch": 3.6161731207289294,
"grad_norm": 1.940301775932312,
"learning_rate": 4.055555555555556e-06,
"loss": 0.1127,
"step": 3175
},
{
"epoch": 3.644646924829157,
"grad_norm": 2.5598275661468506,
"learning_rate": 4.000000000000001e-06,
"loss": 0.106,
"step": 3200
},
{
"epoch": 3.6731207289293852,
"grad_norm": 2.063333511352539,
"learning_rate": 3.944444444444445e-06,
"loss": 0.1077,
"step": 3225
},
{
"epoch": 3.7015945330296125,
"grad_norm": 1.9726035594940186,
"learning_rate": 3.88888888888889e-06,
"loss": 0.117,
"step": 3250
},
{
"epoch": 3.7300683371298406,
"grad_norm": 1.7921056747436523,
"learning_rate": 3.833333333333334e-06,
"loss": 0.1088,
"step": 3275
},
{
"epoch": 3.7585421412300684,
"grad_norm": 2.258916139602661,
"learning_rate": 3.777777777777778e-06,
"loss": 0.1092,
"step": 3300
},
{
"epoch": 3.787015945330296,
"grad_norm": 2.289665937423706,
"learning_rate": 3.7222222222222225e-06,
"loss": 0.108,
"step": 3325
},
{
"epoch": 3.8154897494305238,
"grad_norm": 1.943253755569458,
"learning_rate": 3.6666666666666666e-06,
"loss": 0.111,
"step": 3350
},
{
"epoch": 3.8439635535307515,
"grad_norm": 2.662332773208618,
"learning_rate": 3.6111111111111115e-06,
"loss": 0.1176,
"step": 3375
},
{
"epoch": 3.8724373576309796,
"grad_norm": 2.1651501655578613,
"learning_rate": 3.555555555555556e-06,
"loss": 0.1169,
"step": 3400
},
{
"epoch": 3.9009111617312073,
"grad_norm": 2.1495697498321533,
"learning_rate": 3.5e-06,
"loss": 0.1096,
"step": 3425
},
{
"epoch": 3.929384965831435,
"grad_norm": 2.1727662086486816,
"learning_rate": 3.444444444444445e-06,
"loss": 0.1119,
"step": 3450
},
{
"epoch": 3.9578587699316627,
"grad_norm": 2.4179928302764893,
"learning_rate": 3.3888888888888893e-06,
"loss": 0.1083,
"step": 3475
},
{
"epoch": 3.9863325740318905,
"grad_norm": 2.2201690673828125,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.1097,
"step": 3500
},
{
"epoch": 4.014806378132119,
"grad_norm": 1.7109599113464355,
"learning_rate": 3.277777777777778e-06,
"loss": 0.0834,
"step": 3525
},
{
"epoch": 4.043280182232346,
"grad_norm": 1.5952197313308716,
"learning_rate": 3.2222222222222227e-06,
"loss": 0.0527,
"step": 3550
},
{
"epoch": 4.071753986332574,
"grad_norm": 1.780273675918579,
"learning_rate": 3.1666666666666667e-06,
"loss": 0.0509,
"step": 3575
},
{
"epoch": 4.100227790432802,
"grad_norm": 1.4646574258804321,
"learning_rate": 3.1111111111111116e-06,
"loss": 0.0524,
"step": 3600
},
{
"epoch": 4.128701594533029,
"grad_norm": 1.4881808757781982,
"learning_rate": 3.055555555555556e-06,
"loss": 0.0486,
"step": 3625
},
{
"epoch": 4.157175398633258,
"grad_norm": 1.9114530086517334,
"learning_rate": 3e-06,
"loss": 0.0522,
"step": 3650
},
{
"epoch": 4.185649202733485,
"grad_norm": 1.587689995765686,
"learning_rate": 2.944444444444445e-06,
"loss": 0.0549,
"step": 3675
},
{
"epoch": 4.214123006833713,
"grad_norm": 2.1305880546569824,
"learning_rate": 2.888888888888889e-06,
"loss": 0.0557,
"step": 3700
},
{
"epoch": 4.242596810933941,
"grad_norm": 1.6783413887023926,
"learning_rate": 2.8333333333333335e-06,
"loss": 0.05,
"step": 3725
},
{
"epoch": 4.271070615034168,
"grad_norm": 2.365713119506836,
"learning_rate": 2.7777777777777783e-06,
"loss": 0.0523,
"step": 3750
},
{
"epoch": 4.2995444191343966,
"grad_norm": 1.8531174659729004,
"learning_rate": 2.7222222222222224e-06,
"loss": 0.0555,
"step": 3775
},
{
"epoch": 4.328018223234624,
"grad_norm": 1.5612976551055908,
"learning_rate": 2.666666666666667e-06,
"loss": 0.0562,
"step": 3800
},
{
"epoch": 4.356492027334852,
"grad_norm": 1.8512299060821533,
"learning_rate": 2.6111111111111113e-06,
"loss": 0.0497,
"step": 3825
},
{
"epoch": 4.38496583143508,
"grad_norm": 1.2788305282592773,
"learning_rate": 2.5555555555555557e-06,
"loss": 0.0499,
"step": 3850
},
{
"epoch": 4.413439635535307,
"grad_norm": 2.3926126956939697,
"learning_rate": 2.5e-06,
"loss": 0.0516,
"step": 3875
},
{
"epoch": 4.4419134396355355,
"grad_norm": 1.7549375295639038,
"learning_rate": 2.4444444444444447e-06,
"loss": 0.0527,
"step": 3900
},
{
"epoch": 4.470387243735763,
"grad_norm": 1.603882074356079,
"learning_rate": 2.388888888888889e-06,
"loss": 0.0517,
"step": 3925
},
{
"epoch": 4.498861047835991,
"grad_norm": 1.872842788696289,
"learning_rate": 2.3333333333333336e-06,
"loss": 0.047,
"step": 3950
},
{
"epoch": 4.527334851936219,
"grad_norm": 1.3543041944503784,
"learning_rate": 2.277777777777778e-06,
"loss": 0.0519,
"step": 3975
},
{
"epoch": 4.555808656036446,
"grad_norm": 1.6009862422943115,
"learning_rate": 2.222222222222222e-06,
"loss": 0.0515,
"step": 4000
},
{
"epoch": 4.555808656036446,
"eval_loss": 0.4906230866909027,
"eval_runtime": 2972.4828,
"eval_samples_per_second": 2.362,
"eval_steps_per_second": 0.148,
"eval_wer": 0.28324475152803613,
"step": 4000
},
{
"epoch": 4.5842824601366745,
"grad_norm": 2.2875330448150635,
"learning_rate": 2.166666666666667e-06,
"loss": 0.0494,
"step": 4025
},
{
"epoch": 4.612756264236902,
"grad_norm": 1.5816611051559448,
"learning_rate": 2.1111111111111114e-06,
"loss": 0.0482,
"step": 4050
},
{
"epoch": 4.64123006833713,
"grad_norm": 2.0431058406829834,
"learning_rate": 2.0555555555555555e-06,
"loss": 0.0591,
"step": 4075
},
{
"epoch": 4.669703872437358,
"grad_norm": 1.7807472944259644,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.048,
"step": 4100
},
{
"epoch": 4.698177676537585,
"grad_norm": 1.7173796892166138,
"learning_rate": 1.944444444444445e-06,
"loss": 0.0506,
"step": 4125
},
{
"epoch": 4.7266514806378135,
"grad_norm": 1.8625760078430176,
"learning_rate": 1.888888888888889e-06,
"loss": 0.0493,
"step": 4150
},
{
"epoch": 4.755125284738041,
"grad_norm": 1.7857956886291504,
"learning_rate": 1.8333333333333333e-06,
"loss": 0.0507,
"step": 4175
},
{
"epoch": 4.783599088838269,
"grad_norm": 1.458436369895935,
"learning_rate": 1.777777777777778e-06,
"loss": 0.0517,
"step": 4200
},
{
"epoch": 4.812072892938497,
"grad_norm": 2.7380471229553223,
"learning_rate": 1.7222222222222224e-06,
"loss": 0.0491,
"step": 4225
},
{
"epoch": 4.840546697038724,
"grad_norm": 2.323181390762329,
"learning_rate": 1.6666666666666667e-06,
"loss": 0.0513,
"step": 4250
},
{
"epoch": 4.8690205011389525,
"grad_norm": 2.195849895477295,
"learning_rate": 1.6111111111111113e-06,
"loss": 0.045,
"step": 4275
},
{
"epoch": 4.89749430523918,
"grad_norm": 1.933990478515625,
"learning_rate": 1.5555555555555558e-06,
"loss": 0.0448,
"step": 4300
},
{
"epoch": 4.925968109339408,
"grad_norm": 1.2482439279556274,
"learning_rate": 1.5e-06,
"loss": 0.0471,
"step": 4325
},
{
"epoch": 4.954441913439636,
"grad_norm": 1.8933014869689941,
"learning_rate": 1.4444444444444445e-06,
"loss": 0.0486,
"step": 4350
},
{
"epoch": 4.982915717539863,
"grad_norm": 1.77507746219635,
"learning_rate": 1.3888888888888892e-06,
"loss": 0.0489,
"step": 4375
},
{
"epoch": 5.011389521640091,
"grad_norm": 1.1538957357406616,
"learning_rate": 1.3333333333333334e-06,
"loss": 0.0393,
"step": 4400
},
{
"epoch": 5.039863325740319,
"grad_norm": 0.9963041543960571,
"learning_rate": 1.2777777777777779e-06,
"loss": 0.0226,
"step": 4425
},
{
"epoch": 5.068337129840547,
"grad_norm": 0.8649413585662842,
"learning_rate": 1.2222222222222223e-06,
"loss": 0.0203,
"step": 4450
},
{
"epoch": 5.096810933940774,
"grad_norm": 1.663192629814148,
"learning_rate": 1.1666666666666668e-06,
"loss": 0.0209,
"step": 4475
},
{
"epoch": 5.125284738041002,
"grad_norm": 2.2397210597991943,
"learning_rate": 1.111111111111111e-06,
"loss": 0.0199,
"step": 4500
},
{
"epoch": 5.15375854214123,
"grad_norm": 1.0296459197998047,
"learning_rate": 1.0555555555555557e-06,
"loss": 0.0204,
"step": 4525
},
{
"epoch": 5.182232346241458,
"grad_norm": 1.3559978008270264,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.0195,
"step": 4550
},
{
"epoch": 5.210706150341686,
"grad_norm": 1.0463148355484009,
"learning_rate": 9.444444444444445e-07,
"loss": 0.0243,
"step": 4575
},
{
"epoch": 5.239179954441913,
"grad_norm": 1.8995018005371094,
"learning_rate": 8.88888888888889e-07,
"loss": 0.0204,
"step": 4600
},
{
"epoch": 5.267653758542141,
"grad_norm": 0.8292195796966553,
"learning_rate": 8.333333333333333e-07,
"loss": 0.0185,
"step": 4625
},
{
"epoch": 5.296127562642369,
"grad_norm": 1.4848933219909668,
"learning_rate": 7.777777777777779e-07,
"loss": 0.0206,
"step": 4650
},
{
"epoch": 5.324601366742597,
"grad_norm": 0.9767656922340393,
"learning_rate": 7.222222222222222e-07,
"loss": 0.0202,
"step": 4675
},
{
"epoch": 5.353075170842825,
"grad_norm": 0.8986805081367493,
"learning_rate": 6.666666666666667e-07,
"loss": 0.0185,
"step": 4700
},
{
"epoch": 5.381548974943052,
"grad_norm": 1.233708143234253,
"learning_rate": 6.111111111111112e-07,
"loss": 0.0178,
"step": 4725
},
{
"epoch": 5.41002277904328,
"grad_norm": 1.045877456665039,
"learning_rate": 5.555555555555555e-07,
"loss": 0.0188,
"step": 4750
},
{
"epoch": 5.438496583143508,
"grad_norm": 1.4477061033248901,
"learning_rate": 5.000000000000001e-07,
"loss": 0.0217,
"step": 4775
},
{
"epoch": 5.466970387243736,
"grad_norm": 1.7844326496124268,
"learning_rate": 4.444444444444445e-07,
"loss": 0.0186,
"step": 4800
},
{
"epoch": 5.495444191343964,
"grad_norm": 1.0752902030944824,
"learning_rate": 3.8888888888888895e-07,
"loss": 0.0197,
"step": 4825
},
{
"epoch": 5.523917995444191,
"grad_norm": 1.193974256515503,
"learning_rate": 3.3333333333333335e-07,
"loss": 0.0201,
"step": 4850
},
{
"epoch": 5.552391799544419,
"grad_norm": 0.7838534116744995,
"learning_rate": 2.7777777777777776e-07,
"loss": 0.0202,
"step": 4875
},
{
"epoch": 5.5808656036446465,
"grad_norm": 0.9036350250244141,
"learning_rate": 2.2222222222222224e-07,
"loss": 0.0199,
"step": 4900
},
{
"epoch": 5.609339407744875,
"grad_norm": 0.8704236745834351,
"learning_rate": 1.6666666666666668e-07,
"loss": 0.0198,
"step": 4925
},
{
"epoch": 5.637813211845103,
"grad_norm": 1.007270336151123,
"learning_rate": 1.1111111111111112e-07,
"loss": 0.0222,
"step": 4950
},
{
"epoch": 5.66628701594533,
"grad_norm": 0.9725838899612427,
"learning_rate": 5.555555555555556e-08,
"loss": 0.0196,
"step": 4975
},
{
"epoch": 5.694760820045558,
"grad_norm": 0.970182478427887,
"learning_rate": 0.0,
"loss": 0.0193,
"step": 5000
},
{
"epoch": 5.694760820045558,
"eval_loss": 0.5499717593193054,
"eval_runtime": 2981.2655,
"eval_samples_per_second": 2.355,
"eval_steps_per_second": 0.147,
"eval_wer": 0.2817565771990433,
"step": 5000
},
{
"epoch": 5.694760820045558,
"step": 5000,
"total_flos": 5.43429854134272e+20,
"train_loss": 0.23574846556782722,
"train_runtime": 65192.9802,
"train_samples_per_second": 2.454,
"train_steps_per_second": 0.077
}
],
"logging_steps": 25,
"max_steps": 5000,
"num_input_tokens_seen": 0,
"num_train_epochs": 6,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.43429854134272e+20,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}