{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.3137151865475565, "eval_steps": 1000, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.006568575932737782, "grad_norm": 4.938570976257324, "learning_rate": 5.000000000000001e-07, "loss": 1.1658, "step": 25 }, { "epoch": 0.013137151865475564, "grad_norm": 3.4416959285736084, "learning_rate": 1.0000000000000002e-06, "loss": 0.9316, "step": 50 }, { "epoch": 0.01970572779821335, "grad_norm": 3.3691599369049072, "learning_rate": 1.5e-06, "loss": 0.753, "step": 75 }, { "epoch": 0.02627430373095113, "grad_norm": 3.2769742012023926, "learning_rate": 2.0000000000000003e-06, "loss": 0.6576, "step": 100 }, { "epoch": 0.032842879663688915, "grad_norm": 2.5112249851226807, "learning_rate": 2.5e-06, "loss": 0.6188, "step": 125 }, { "epoch": 0.0394114555964267, "grad_norm": 3.0561046600341797, "learning_rate": 3e-06, "loss": 0.6021, "step": 150 }, { "epoch": 0.045980031529164474, "grad_norm": 2.563472032546997, "learning_rate": 3.5e-06, "loss": 0.564, "step": 175 }, { "epoch": 0.05254860746190226, "grad_norm": 2.6831862926483154, "learning_rate": 4.000000000000001e-06, "loss": 0.5314, "step": 200 }, { "epoch": 0.05911718339464004, "grad_norm": 2.383794069290161, "learning_rate": 4.5e-06, "loss": 0.5376, "step": 225 }, { "epoch": 0.06568575932737783, "grad_norm": 2.8369593620300293, "learning_rate": 5e-06, "loss": 0.5404, "step": 250 }, { "epoch": 0.07225433526011561, "grad_norm": 2.641814708709717, "learning_rate": 5.500000000000001e-06, "loss": 0.5014, "step": 275 }, { "epoch": 0.0788229111928534, "grad_norm": 2.549522638320923, "learning_rate": 6e-06, "loss": 0.4978, "step": 300 }, { "epoch": 0.08539148712559118, "grad_norm": 2.770048141479492, "learning_rate": 6.5000000000000004e-06, "loss": 0.5135, "step": 325 }, { "epoch": 0.09196006305832895, "grad_norm": 3.170330286026001, "learning_rate": 7e-06, "loss": 0.504, "step": 350 }, { "epoch": 0.09852863899106673, "grad_norm": 2.712066650390625, "learning_rate": 7.500000000000001e-06, "loss": 0.451, "step": 375 }, { "epoch": 0.10509721492380451, "grad_norm": 2.3632166385650635, "learning_rate": 8.000000000000001e-06, "loss": 0.4533, "step": 400 }, { "epoch": 0.1116657908565423, "grad_norm": 2.728026866912842, "learning_rate": 8.5e-06, "loss": 0.4732, "step": 425 }, { "epoch": 0.11823436678928008, "grad_norm": 2.7299931049346924, "learning_rate": 9e-06, "loss": 0.4255, "step": 450 }, { "epoch": 0.12480294272201786, "grad_norm": 2.4892563819885254, "learning_rate": 9.5e-06, "loss": 0.4254, "step": 475 }, { "epoch": 0.13137151865475566, "grad_norm": 2.3553197383880615, "learning_rate": 1e-05, "loss": 0.4218, "step": 500 }, { "epoch": 0.13794009458749343, "grad_norm": 2.1275930404663086, "learning_rate": 9.944444444444445e-06, "loss": 0.4159, "step": 525 }, { "epoch": 0.14450867052023122, "grad_norm": 2.166019916534424, "learning_rate": 9.88888888888889e-06, "loss": 0.4051, "step": 550 }, { "epoch": 0.151077246452969, "grad_norm": 2.5494415760040283, "learning_rate": 9.833333333333333e-06, "loss": 0.399, "step": 575 }, { "epoch": 0.1576458223857068, "grad_norm": 2.482174873352051, "learning_rate": 9.777777777777779e-06, "loss": 0.4013, "step": 600 }, { "epoch": 0.16421439831844456, "grad_norm": 2.3230655193328857, "learning_rate": 9.722222222222223e-06, "loss": 0.3736, "step": 625 }, { "epoch": 0.17078297425118236, "grad_norm": 2.1272637844085693, "learning_rate": 9.666666666666667e-06, "loss": 0.3715, "step": 650 }, { "epoch": 0.17735155018392013, "grad_norm": 2.2144370079040527, "learning_rate": 9.611111111111112e-06, "loss": 0.389, "step": 675 }, { "epoch": 0.1839201261166579, "grad_norm": 2.156562566757202, "learning_rate": 9.555555555555556e-06, "loss": 0.3779, "step": 700 }, { "epoch": 0.1904887020493957, "grad_norm": 1.9418479204177856, "learning_rate": 9.5e-06, "loss": 0.3624, "step": 725 }, { "epoch": 0.19705727798213346, "grad_norm": 1.9420461654663086, "learning_rate": 9.444444444444445e-06, "loss": 0.3726, "step": 750 }, { "epoch": 0.20362585391487126, "grad_norm": 1.929376482963562, "learning_rate": 9.38888888888889e-06, "loss": 0.3501, "step": 775 }, { "epoch": 0.21019442984760903, "grad_norm": 2.257051467895508, "learning_rate": 9.333333333333334e-06, "loss": 0.3755, "step": 800 }, { "epoch": 0.21676300578034682, "grad_norm": 2.102417230606079, "learning_rate": 9.277777777777778e-06, "loss": 0.3671, "step": 825 }, { "epoch": 0.2233315817130846, "grad_norm": 2.096370220184326, "learning_rate": 9.222222222222224e-06, "loss": 0.3397, "step": 850 }, { "epoch": 0.2299001576458224, "grad_norm": 2.0005109310150146, "learning_rate": 9.166666666666666e-06, "loss": 0.3467, "step": 875 }, { "epoch": 0.23646873357856016, "grad_norm": 2.1975176334381104, "learning_rate": 9.111111111111112e-06, "loss": 0.3341, "step": 900 }, { "epoch": 0.24303730951129796, "grad_norm": 2.0950164794921875, "learning_rate": 9.055555555555556e-06, "loss": 0.3398, "step": 925 }, { "epoch": 0.24960588544403572, "grad_norm": 2.0137791633605957, "learning_rate": 9e-06, "loss": 0.3183, "step": 950 }, { "epoch": 0.2561744613767735, "grad_norm": 1.97981595993042, "learning_rate": 8.944444444444446e-06, "loss": 0.3653, "step": 975 }, { "epoch": 0.2627430373095113, "grad_norm": 1.9213696718215942, "learning_rate": 8.888888888888888e-06, "loss": 0.3429, "step": 1000 }, { "epoch": 0.2627430373095113, "eval_loss": 0.3295910656452179, "eval_runtime": 594.427, "eval_samples_per_second": 2.052, "eval_steps_per_second": 0.13, "eval_wer": 0.2299665865525529, "step": 1000 }, { "epoch": 0.26931161324224906, "grad_norm": 2.122868299484253, "learning_rate": 8.833333333333334e-06, "loss": 0.318, "step": 1025 }, { "epoch": 0.27588018917498686, "grad_norm": 2.259373426437378, "learning_rate": 8.777777777777778e-06, "loss": 0.3364, "step": 1050 }, { "epoch": 0.28244876510772465, "grad_norm": 1.978367567062378, "learning_rate": 8.722222222222224e-06, "loss": 0.3125, "step": 1075 }, { "epoch": 0.28901734104046245, "grad_norm": 1.8453904390335083, "learning_rate": 8.666666666666668e-06, "loss": 0.3009, "step": 1100 }, { "epoch": 0.2955859169732002, "grad_norm": 2.002053737640381, "learning_rate": 8.611111111111112e-06, "loss": 0.3087, "step": 1125 }, { "epoch": 0.302154492905938, "grad_norm": 2.080690860748291, "learning_rate": 8.555555555555556e-06, "loss": 0.3188, "step": 1150 }, { "epoch": 0.3087230688386758, "grad_norm": 1.6879727840423584, "learning_rate": 8.5e-06, "loss": 0.3043, "step": 1175 }, { "epoch": 0.3152916447714136, "grad_norm": 2.2736353874206543, "learning_rate": 8.444444444444446e-06, "loss": 0.3165, "step": 1200 }, { "epoch": 0.3218602207041513, "grad_norm": 1.8325968980789185, "learning_rate": 8.38888888888889e-06, "loss": 0.3236, "step": 1225 }, { "epoch": 0.3284287966368891, "grad_norm": 1.8235770463943481, "learning_rate": 8.333333333333334e-06, "loss": 0.3057, "step": 1250 }, { "epoch": 0.3349973725696269, "grad_norm": 2.09566068649292, "learning_rate": 8.277777777777778e-06, "loss": 0.2845, "step": 1275 }, { "epoch": 0.3415659485023647, "grad_norm": 1.9097365140914917, "learning_rate": 8.222222222222222e-06, "loss": 0.2833, "step": 1300 }, { "epoch": 0.34813452443510245, "grad_norm": 1.9709367752075195, "learning_rate": 8.166666666666668e-06, "loss": 0.294, "step": 1325 }, { "epoch": 0.35470310036784025, "grad_norm": 1.9940749406814575, "learning_rate": 8.111111111111112e-06, "loss": 0.2789, "step": 1350 }, { "epoch": 0.36127167630057805, "grad_norm": 2.122657299041748, "learning_rate": 8.055555555555557e-06, "loss": 0.3026, "step": 1375 }, { "epoch": 0.3678402522333158, "grad_norm": 1.742019772529602, "learning_rate": 8.000000000000001e-06, "loss": 0.2843, "step": 1400 }, { "epoch": 0.3744088281660536, "grad_norm": 1.8771026134490967, "learning_rate": 7.944444444444445e-06, "loss": 0.2937, "step": 1425 }, { "epoch": 0.3809774040987914, "grad_norm": 1.617311954498291, "learning_rate": 7.88888888888889e-06, "loss": 0.2749, "step": 1450 }, { "epoch": 0.3875459800315292, "grad_norm": 1.902928113937378, "learning_rate": 7.833333333333333e-06, "loss": 0.2929, "step": 1475 }, { "epoch": 0.3941145559642669, "grad_norm": 1.9772895574569702, "learning_rate": 7.77777777777778e-06, "loss": 0.2699, "step": 1500 }, { "epoch": 0.4006831318970047, "grad_norm": 1.7294281721115112, "learning_rate": 7.722222222222223e-06, "loss": 0.288, "step": 1525 }, { "epoch": 0.4072517078297425, "grad_norm": 1.6163759231567383, "learning_rate": 7.666666666666667e-06, "loss": 0.2919, "step": 1550 }, { "epoch": 0.4138202837624803, "grad_norm": 1.9823168516159058, "learning_rate": 7.611111111111111e-06, "loss": 0.2691, "step": 1575 }, { "epoch": 0.42038885969521805, "grad_norm": 2.085510730743408, "learning_rate": 7.555555555555556e-06, "loss": 0.2807, "step": 1600 }, { "epoch": 0.42695743562795585, "grad_norm": 2.1536219120025635, "learning_rate": 7.500000000000001e-06, "loss": 0.2977, "step": 1625 }, { "epoch": 0.43352601156069365, "grad_norm": 2.23018741607666, "learning_rate": 7.444444444444445e-06, "loss": 0.291, "step": 1650 }, { "epoch": 0.44009458749343144, "grad_norm": 1.8097656965255737, "learning_rate": 7.38888888888889e-06, "loss": 0.3125, "step": 1675 }, { "epoch": 0.4466631634261692, "grad_norm": 1.557788372039795, "learning_rate": 7.333333333333333e-06, "loss": 0.2743, "step": 1700 }, { "epoch": 0.453231739358907, "grad_norm": 1.677343726158142, "learning_rate": 7.277777777777778e-06, "loss": 0.2568, "step": 1725 }, { "epoch": 0.4598003152916448, "grad_norm": 1.7231301069259644, "learning_rate": 7.222222222222223e-06, "loss": 0.2701, "step": 1750 }, { "epoch": 0.4663688912243826, "grad_norm": 1.632379412651062, "learning_rate": 7.166666666666667e-06, "loss": 0.2795, "step": 1775 }, { "epoch": 0.4729374671571203, "grad_norm": 2.042989730834961, "learning_rate": 7.111111111111112e-06, "loss": 0.2714, "step": 1800 }, { "epoch": 0.4795060430898581, "grad_norm": 1.8419965505599976, "learning_rate": 7.055555555555557e-06, "loss": 0.2638, "step": 1825 }, { "epoch": 0.4860746190225959, "grad_norm": 1.7692322731018066, "learning_rate": 7e-06, "loss": 0.2631, "step": 1850 }, { "epoch": 0.4926431949553337, "grad_norm": 1.8727885484695435, "learning_rate": 6.944444444444445e-06, "loss": 0.2647, "step": 1875 }, { "epoch": 0.49921177088807145, "grad_norm": 1.5731993913650513, "learning_rate": 6.88888888888889e-06, "loss": 0.2765, "step": 1900 }, { "epoch": 0.5057803468208093, "grad_norm": 1.9374446868896484, "learning_rate": 6.833333333333334e-06, "loss": 0.2672, "step": 1925 }, { "epoch": 0.512348922753547, "grad_norm": 1.702355980873108, "learning_rate": 6.777777777777779e-06, "loss": 0.2574, "step": 1950 }, { "epoch": 0.5189174986862848, "grad_norm": 1.539027214050293, "learning_rate": 6.7222222222222235e-06, "loss": 0.2651, "step": 1975 }, { "epoch": 0.5254860746190226, "grad_norm": 1.8967093229293823, "learning_rate": 6.666666666666667e-06, "loss": 0.2636, "step": 2000 }, { "epoch": 0.5254860746190226, "eval_loss": 0.2615206837654114, "eval_runtime": 532.8353, "eval_samples_per_second": 2.29, "eval_steps_per_second": 0.145, "eval_wer": 0.1925669734451023, "step": 2000 }, { "epoch": 0.5320546505517604, "grad_norm": 1.688244342803955, "learning_rate": 6.6111111111111115e-06, "loss": 0.2683, "step": 2025 }, { "epoch": 0.5386232264844981, "grad_norm": 1.8095322847366333, "learning_rate": 6.555555555555556e-06, "loss": 0.2707, "step": 2050 }, { "epoch": 0.545191802417236, "grad_norm": 1.8945379257202148, "learning_rate": 6.5000000000000004e-06, "loss": 0.2535, "step": 2075 }, { "epoch": 0.5517603783499737, "grad_norm": 1.9910788536071777, "learning_rate": 6.444444444444445e-06, "loss": 0.2486, "step": 2100 }, { "epoch": 0.5583289542827115, "grad_norm": 1.3568848371505737, "learning_rate": 6.3888888888888885e-06, "loss": 0.2755, "step": 2125 }, { "epoch": 0.5648975302154493, "grad_norm": 1.7561497688293457, "learning_rate": 6.333333333333333e-06, "loss": 0.2679, "step": 2150 }, { "epoch": 0.571466106148187, "grad_norm": 1.6052385568618774, "learning_rate": 6.277777777777778e-06, "loss": 0.2497, "step": 2175 }, { "epoch": 0.5780346820809249, "grad_norm": 1.5882091522216797, "learning_rate": 6.222222222222223e-06, "loss": 0.2673, "step": 2200 }, { "epoch": 0.5846032580136626, "grad_norm": 1.7540193796157837, "learning_rate": 6.166666666666667e-06, "loss": 0.2335, "step": 2225 }, { "epoch": 0.5911718339464004, "grad_norm": 1.4882384538650513, "learning_rate": 6.111111111111112e-06, "loss": 0.2555, "step": 2250 }, { "epoch": 0.5977404098791382, "grad_norm": 2.00600266456604, "learning_rate": 6.055555555555555e-06, "loss": 0.2687, "step": 2275 }, { "epoch": 0.604308985811876, "grad_norm": 1.4347535371780396, "learning_rate": 6e-06, "loss": 0.2681, "step": 2300 }, { "epoch": 0.6108775617446137, "grad_norm": 1.8053362369537354, "learning_rate": 5.944444444444445e-06, "loss": 0.2568, "step": 2325 }, { "epoch": 0.6174461376773516, "grad_norm": 1.657359004020691, "learning_rate": 5.88888888888889e-06, "loss": 0.246, "step": 2350 }, { "epoch": 0.6240147136100893, "grad_norm": 1.6917238235473633, "learning_rate": 5.833333333333334e-06, "loss": 0.2333, "step": 2375 }, { "epoch": 0.6305832895428272, "grad_norm": 2.0534191131591797, "learning_rate": 5.777777777777778e-06, "loss": 0.2557, "step": 2400 }, { "epoch": 0.6371518654755649, "grad_norm": 1.6997853517532349, "learning_rate": 5.722222222222222e-06, "loss": 0.2888, "step": 2425 }, { "epoch": 0.6437204414083026, "grad_norm": 1.7183536291122437, "learning_rate": 5.666666666666667e-06, "loss": 0.2364, "step": 2450 }, { "epoch": 0.6502890173410405, "grad_norm": 1.6452535390853882, "learning_rate": 5.611111111111112e-06, "loss": 0.2395, "step": 2475 }, { "epoch": 0.6568575932737782, "grad_norm": 1.7632161378860474, "learning_rate": 5.555555555555557e-06, "loss": 0.2547, "step": 2500 }, { "epoch": 0.663426169206516, "grad_norm": 1.9056440591812134, "learning_rate": 5.500000000000001e-06, "loss": 0.249, "step": 2525 }, { "epoch": 0.6699947451392538, "grad_norm": 2.083576202392578, "learning_rate": 5.444444444444445e-06, "loss": 0.2596, "step": 2550 }, { "epoch": 0.6765633210719916, "grad_norm": 2.1211483478546143, "learning_rate": 5.388888888888889e-06, "loss": 0.2435, "step": 2575 }, { "epoch": 0.6831318970047294, "grad_norm": 1.599048376083374, "learning_rate": 5.333333333333334e-06, "loss": 0.2376, "step": 2600 }, { "epoch": 0.6897004729374672, "grad_norm": 1.6461100578308105, "learning_rate": 5.2777777777777785e-06, "loss": 0.2457, "step": 2625 }, { "epoch": 0.6962690488702049, "grad_norm": 1.660069227218628, "learning_rate": 5.2222222222222226e-06, "loss": 0.2354, "step": 2650 }, { "epoch": 0.7028376248029428, "grad_norm": 2.221201181411743, "learning_rate": 5.1666666666666675e-06, "loss": 0.2794, "step": 2675 }, { "epoch": 0.7094062007356805, "grad_norm": 2.4395995140075684, "learning_rate": 5.1111111111111115e-06, "loss": 0.2499, "step": 2700 }, { "epoch": 0.7159747766684182, "grad_norm": 1.4763926267623901, "learning_rate": 5.0555555555555555e-06, "loss": 0.2783, "step": 2725 }, { "epoch": 0.7225433526011561, "grad_norm": 1.6573847532272339, "learning_rate": 5e-06, "loss": 0.2349, "step": 2750 }, { "epoch": 0.7291119285338938, "grad_norm": 1.8241550922393799, "learning_rate": 4.944444444444445e-06, "loss": 0.2449, "step": 2775 }, { "epoch": 0.7356805044666316, "grad_norm": 1.6332080364227295, "learning_rate": 4.888888888888889e-06, "loss": 0.2413, "step": 2800 }, { "epoch": 0.7422490803993694, "grad_norm": 1.626541256904602, "learning_rate": 4.833333333333333e-06, "loss": 0.2499, "step": 2825 }, { "epoch": 0.7488176563321072, "grad_norm": 1.5557328462600708, "learning_rate": 4.777777777777778e-06, "loss": 0.2534, "step": 2850 }, { "epoch": 0.755386232264845, "grad_norm": 1.932283878326416, "learning_rate": 4.722222222222222e-06, "loss": 0.2379, "step": 2875 }, { "epoch": 0.7619548081975828, "grad_norm": 1.9064332246780396, "learning_rate": 4.666666666666667e-06, "loss": 0.2323, "step": 2900 }, { "epoch": 0.7685233841303205, "grad_norm": 1.5515189170837402, "learning_rate": 4.611111111111112e-06, "loss": 0.2328, "step": 2925 }, { "epoch": 0.7750919600630584, "grad_norm": 2.1047863960266113, "learning_rate": 4.555555555555556e-06, "loss": 0.248, "step": 2950 }, { "epoch": 0.7816605359957961, "grad_norm": 1.8366841077804565, "learning_rate": 4.5e-06, "loss": 0.2584, "step": 2975 }, { "epoch": 0.7882291119285338, "grad_norm": 1.949874997138977, "learning_rate": 4.444444444444444e-06, "loss": 0.2316, "step": 3000 }, { "epoch": 0.7882291119285338, "eval_loss": 0.23398828506469727, "eval_runtime": 533.7623, "eval_samples_per_second": 2.286, "eval_steps_per_second": 0.144, "eval_wer": 0.1797877952986693, "step": 3000 }, { "epoch": 0.7947976878612717, "grad_norm": 2.0107836723327637, "learning_rate": 4.388888888888889e-06, "loss": 0.2421, "step": 3025 }, { "epoch": 0.8013662637940094, "grad_norm": 1.959560751914978, "learning_rate": 4.333333333333334e-06, "loss": 0.2319, "step": 3050 }, { "epoch": 0.8079348397267473, "grad_norm": 1.6217644214630127, "learning_rate": 4.277777777777778e-06, "loss": 0.2383, "step": 3075 }, { "epoch": 0.814503415659485, "grad_norm": 2.180353879928589, "learning_rate": 4.222222222222223e-06, "loss": 0.2374, "step": 3100 }, { "epoch": 0.8210719915922228, "grad_norm": 1.7979501485824585, "learning_rate": 4.166666666666667e-06, "loss": 0.2437, "step": 3125 }, { "epoch": 0.8276405675249606, "grad_norm": 1.6101832389831543, "learning_rate": 4.111111111111111e-06, "loss": 0.2142, "step": 3150 }, { "epoch": 0.8342091434576984, "grad_norm": 2.010740280151367, "learning_rate": 4.055555555555556e-06, "loss": 0.2489, "step": 3175 }, { "epoch": 0.8407777193904361, "grad_norm": 1.62699556350708, "learning_rate": 4.000000000000001e-06, "loss": 0.2362, "step": 3200 }, { "epoch": 0.847346295323174, "grad_norm": 1.6320332288742065, "learning_rate": 3.944444444444445e-06, "loss": 0.2134, "step": 3225 }, { "epoch": 0.8539148712559117, "grad_norm": 1.5167447328567505, "learning_rate": 3.88888888888889e-06, "loss": 0.219, "step": 3250 }, { "epoch": 0.8604834471886496, "grad_norm": 1.8277373313903809, "learning_rate": 3.833333333333334e-06, "loss": 0.2593, "step": 3275 }, { "epoch": 0.8670520231213873, "grad_norm": 1.6318010091781616, "learning_rate": 3.777777777777778e-06, "loss": 0.2151, "step": 3300 }, { "epoch": 0.873620599054125, "grad_norm": 1.9252204895019531, "learning_rate": 3.7222222222222225e-06, "loss": 0.2509, "step": 3325 }, { "epoch": 0.8801891749868629, "grad_norm": 2.156442880630493, "learning_rate": 3.6666666666666666e-06, "loss": 0.2407, "step": 3350 }, { "epoch": 0.8867577509196006, "grad_norm": 1.9615235328674316, "learning_rate": 3.6111111111111115e-06, "loss": 0.2299, "step": 3375 }, { "epoch": 0.8933263268523384, "grad_norm": 1.985635757446289, "learning_rate": 3.555555555555556e-06, "loss": 0.2451, "step": 3400 }, { "epoch": 0.8998949027850762, "grad_norm": 1.5059348344802856, "learning_rate": 3.5e-06, "loss": 0.2389, "step": 3425 }, { "epoch": 0.906463478717814, "grad_norm": 1.8141510486602783, "learning_rate": 3.444444444444445e-06, "loss": 0.2412, "step": 3450 }, { "epoch": 0.9130320546505517, "grad_norm": 1.7190760374069214, "learning_rate": 3.3888888888888893e-06, "loss": 0.217, "step": 3475 }, { "epoch": 0.9196006305832896, "grad_norm": 1.7403192520141602, "learning_rate": 3.3333333333333333e-06, "loss": 0.2386, "step": 3500 }, { "epoch": 0.9261692065160273, "grad_norm": 1.5826787948608398, "learning_rate": 3.277777777777778e-06, "loss": 0.1965, "step": 3525 }, { "epoch": 0.9327377824487652, "grad_norm": 1.8713033199310303, "learning_rate": 3.2222222222222227e-06, "loss": 0.2205, "step": 3550 }, { "epoch": 0.9393063583815029, "grad_norm": 1.6279629468917847, "learning_rate": 3.1666666666666667e-06, "loss": 0.2031, "step": 3575 }, { "epoch": 0.9458749343142406, "grad_norm": 1.5993796586990356, "learning_rate": 3.1111111111111116e-06, "loss": 0.218, "step": 3600 }, { "epoch": 0.9524435102469785, "grad_norm": 1.8748666048049927, "learning_rate": 3.055555555555556e-06, "loss": 0.2198, "step": 3625 }, { "epoch": 0.9590120861797162, "grad_norm": 2.0783231258392334, "learning_rate": 3e-06, "loss": 0.2285, "step": 3650 }, { "epoch": 0.965580662112454, "grad_norm": 1.6366628408432007, "learning_rate": 2.944444444444445e-06, "loss": 0.2244, "step": 3675 }, { "epoch": 0.9721492380451918, "grad_norm": 1.4321212768554688, "learning_rate": 2.888888888888889e-06, "loss": 0.2129, "step": 3700 }, { "epoch": 0.9787178139779296, "grad_norm": 1.9172074794769287, "learning_rate": 2.8333333333333335e-06, "loss": 0.2308, "step": 3725 }, { "epoch": 0.9852863899106674, "grad_norm": 1.601163387298584, "learning_rate": 2.7777777777777783e-06, "loss": 0.2249, "step": 3750 }, { "epoch": 0.9918549658434052, "grad_norm": 1.920453429222107, "learning_rate": 2.7222222222222224e-06, "loss": 0.241, "step": 3775 }, { "epoch": 0.9984235417761429, "grad_norm": 1.421399712562561, "learning_rate": 2.666666666666667e-06, "loss": 0.245, "step": 3800 }, { "epoch": 1.0049921177088808, "grad_norm": 1.2747637033462524, "learning_rate": 2.6111111111111113e-06, "loss": 0.183, "step": 3825 }, { "epoch": 1.0115606936416186, "grad_norm": 1.276963472366333, "learning_rate": 2.5555555555555557e-06, "loss": 0.1669, "step": 3850 }, { "epoch": 1.0181292695743562, "grad_norm": 1.6604379415512085, "learning_rate": 2.5e-06, "loss": 0.1685, "step": 3875 }, { "epoch": 1.024697845507094, "grad_norm": 1.4884883165359497, "learning_rate": 2.4444444444444447e-06, "loss": 0.1837, "step": 3900 }, { "epoch": 1.031266421439832, "grad_norm": 1.2239917516708374, "learning_rate": 2.388888888888889e-06, "loss": 0.161, "step": 3925 }, { "epoch": 1.0378349973725696, "grad_norm": 1.7419662475585938, "learning_rate": 2.3333333333333336e-06, "loss": 0.1903, "step": 3950 }, { "epoch": 1.0444035733053074, "grad_norm": 1.5548486709594727, "learning_rate": 2.277777777777778e-06, "loss": 0.159, "step": 3975 }, { "epoch": 1.0509721492380453, "grad_norm": 1.7805577516555786, "learning_rate": 2.222222222222222e-06, "loss": 0.1779, "step": 4000 }, { "epoch": 1.0509721492380453, "eval_loss": 0.21787120401859283, "eval_runtime": 531.6511, "eval_samples_per_second": 2.295, "eval_steps_per_second": 0.145, "eval_wer": 0.16243625065947595, "step": 4000 }, { "epoch": 1.057540725170783, "grad_norm": 1.646399974822998, "learning_rate": 2.166666666666667e-06, "loss": 0.1891, "step": 4025 }, { "epoch": 1.0641093011035208, "grad_norm": 2.159637928009033, "learning_rate": 2.1111111111111114e-06, "loss": 0.1712, "step": 4050 }, { "epoch": 1.0706778770362586, "grad_norm": 1.5517977476119995, "learning_rate": 2.0555555555555555e-06, "loss": 0.1582, "step": 4075 }, { "epoch": 1.0772464529689962, "grad_norm": 1.410233974456787, "learning_rate": 2.0000000000000003e-06, "loss": 0.1828, "step": 4100 }, { "epoch": 1.083815028901734, "grad_norm": 1.6741544008255005, "learning_rate": 1.944444444444445e-06, "loss": 0.1597, "step": 4125 }, { "epoch": 1.090383604834472, "grad_norm": 1.508376121520996, "learning_rate": 1.888888888888889e-06, "loss": 0.166, "step": 4150 }, { "epoch": 1.0969521807672096, "grad_norm": 1.4631690979003906, "learning_rate": 1.8333333333333333e-06, "loss": 0.159, "step": 4175 }, { "epoch": 1.1035207566999474, "grad_norm": 1.7129231691360474, "learning_rate": 1.777777777777778e-06, "loss": 0.1682, "step": 4200 }, { "epoch": 1.1100893326326853, "grad_norm": 1.7223049402236938, "learning_rate": 1.7222222222222224e-06, "loss": 0.1865, "step": 4225 }, { "epoch": 1.116657908565423, "grad_norm": 1.8834460973739624, "learning_rate": 1.6666666666666667e-06, "loss": 0.1753, "step": 4250 }, { "epoch": 1.1232264844981608, "grad_norm": 1.6631660461425781, "learning_rate": 1.6111111111111113e-06, "loss": 0.1585, "step": 4275 }, { "epoch": 1.1297950604308986, "grad_norm": 1.6180058717727661, "learning_rate": 1.5555555555555558e-06, "loss": 0.1839, "step": 4300 }, { "epoch": 1.1363636363636362, "grad_norm": 1.635799527168274, "learning_rate": 1.5e-06, "loss": 0.1685, "step": 4325 }, { "epoch": 1.142932212296374, "grad_norm": 1.4893510341644287, "learning_rate": 1.4444444444444445e-06, "loss": 0.1678, "step": 4350 }, { "epoch": 1.149500788229112, "grad_norm": 1.4803720712661743, "learning_rate": 1.3888888888888892e-06, "loss": 0.1607, "step": 4375 }, { "epoch": 1.1560693641618498, "grad_norm": 1.4839766025543213, "learning_rate": 1.3333333333333334e-06, "loss": 0.187, "step": 4400 }, { "epoch": 1.1626379400945874, "grad_norm": 1.518723487854004, "learning_rate": 1.2777777777777779e-06, "loss": 0.1628, "step": 4425 }, { "epoch": 1.1692065160273253, "grad_norm": 1.740968942642212, "learning_rate": 1.2222222222222223e-06, "loss": 0.1521, "step": 4450 }, { "epoch": 1.1757750919600631, "grad_norm": 1.941607117652893, "learning_rate": 1.1666666666666668e-06, "loss": 0.1558, "step": 4475 }, { "epoch": 1.1823436678928008, "grad_norm": 1.47350013256073, "learning_rate": 1.111111111111111e-06, "loss": 0.1758, "step": 4500 }, { "epoch": 1.1889122438255386, "grad_norm": 1.4565749168395996, "learning_rate": 1.0555555555555557e-06, "loss": 0.1804, "step": 4525 }, { "epoch": 1.1954808197582765, "grad_norm": 1.6238913536071777, "learning_rate": 1.0000000000000002e-06, "loss": 0.1596, "step": 4550 }, { "epoch": 1.202049395691014, "grad_norm": 1.6127305030822754, "learning_rate": 9.444444444444445e-07, "loss": 0.1801, "step": 4575 }, { "epoch": 1.208617971623752, "grad_norm": 1.5954357385635376, "learning_rate": 8.88888888888889e-07, "loss": 0.1616, "step": 4600 }, { "epoch": 1.2151865475564898, "grad_norm": 1.5588889122009277, "learning_rate": 8.333333333333333e-07, "loss": 0.1727, "step": 4625 }, { "epoch": 1.2217551234892277, "grad_norm": 1.2232043743133545, "learning_rate": 7.777777777777779e-07, "loss": 0.1617, "step": 4650 }, { "epoch": 1.2283236994219653, "grad_norm": 1.3071914911270142, "learning_rate": 7.222222222222222e-07, "loss": 0.1623, "step": 4675 }, { "epoch": 1.2348922753547031, "grad_norm": 1.1346869468688965, "learning_rate": 6.666666666666667e-07, "loss": 0.155, "step": 4700 }, { "epoch": 1.241460851287441, "grad_norm": 1.572501540184021, "learning_rate": 6.111111111111112e-07, "loss": 0.1742, "step": 4725 }, { "epoch": 1.2480294272201786, "grad_norm": 1.6292036771774292, "learning_rate": 5.555555555555555e-07, "loss": 0.1535, "step": 4750 }, { "epoch": 1.2545980031529165, "grad_norm": 1.8764897584915161, "learning_rate": 5.000000000000001e-07, "loss": 0.1564, "step": 4775 }, { "epoch": 1.2611665790856543, "grad_norm": 1.7719995975494385, "learning_rate": 4.444444444444445e-07, "loss": 0.1779, "step": 4800 }, { "epoch": 1.267735155018392, "grad_norm": 1.4942494630813599, "learning_rate": 3.8888888888888895e-07, "loss": 0.1743, "step": 4825 }, { "epoch": 1.2743037309511298, "grad_norm": 1.6207414865493774, "learning_rate": 3.3333333333333335e-07, "loss": 0.1731, "step": 4850 }, { "epoch": 1.2808723068838677, "grad_norm": 1.4872610569000244, "learning_rate": 2.7777777777777776e-07, "loss": 0.1708, "step": 4875 }, { "epoch": 1.2874408828166053, "grad_norm": 1.8898082971572876, "learning_rate": 2.2222222222222224e-07, "loss": 0.1819, "step": 4900 }, { "epoch": 1.2940094587493431, "grad_norm": 1.4731193780899048, "learning_rate": 1.6666666666666668e-07, "loss": 0.1714, "step": 4925 }, { "epoch": 1.300578034682081, "grad_norm": 1.4309362173080444, "learning_rate": 1.1111111111111112e-07, "loss": 0.1668, "step": 4950 }, { "epoch": 1.3071466106148186, "grad_norm": 1.6482913494110107, "learning_rate": 5.555555555555556e-08, "loss": 0.1928, "step": 4975 }, { "epoch": 1.3137151865475565, "grad_norm": 1.3571888208389282, "learning_rate": 0.0, "loss": 0.1626, "step": 5000 }, { "epoch": 1.3137151865475565, "eval_loss": 0.21081987023353577, "eval_runtime": 712.9205, "eval_samples_per_second": 1.711, "eval_steps_per_second": 0.108, "eval_wer": 0.1561639017527405, "step": 5000 }, { "epoch": 1.3137151865475565, "step": 5000, "total_flos": 5.435725490631475e+20, "train_loss": 0.2802161669254303, "train_runtime": 36290.7707, "train_samples_per_second": 4.409, "train_steps_per_second": 0.138 } ], "logging_steps": 25, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.435725490631475e+20, "train_batch_size": 16, "trial_name": null, "trial_params": null }