{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.994236311239193, "eval_steps": 20, "global_step": 346, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.011527377521613832, "grad_norm": 0.0, "learning_rate": 0.0, "loss": 0.7662, "step": 2 }, { "epoch": 0.023054755043227664, "grad_norm": 0.0, "learning_rate": 0.0, "loss": 0.7773, "step": 4 }, { "epoch": 0.0345821325648415, "grad_norm": 0.0, "learning_rate": 0.0, "loss": 0.7593, "step": 6 }, { "epoch": 0.04610951008645533, "grad_norm": 0.0, "learning_rate": 0.0, "loss": 0.795, "step": 8 }, { "epoch": 0.05763688760806916, "grad_norm": 0.0, "learning_rate": 0.0, "loss": 0.7765, "step": 10 }, { "epoch": 0.069164265129683, "grad_norm": 0.0, "learning_rate": 0.0, "loss": 0.7819, "step": 12 }, { "epoch": 0.08069164265129683, "grad_norm": 1.0932377576828003, "learning_rate": 2.5e-06, "loss": 0.7426, "step": 14 }, { "epoch": 0.09221902017291066, "grad_norm": 1.0143316984176636, "learning_rate": 7.5e-06, "loss": 0.7667, "step": 16 }, { "epoch": 0.1037463976945245, "grad_norm": 0.5817523002624512, "learning_rate": 1.25e-05, "loss": 0.6819, "step": 18 }, { "epoch": 0.11527377521613832, "grad_norm": 0.7819337248802185, "learning_rate": 1.5e-05, "loss": 0.6171, "step": 20 }, { "epoch": 0.11527377521613832, "eval_loss": 0.5882565379142761, "eval_runtime": 219.1309, "eval_samples_per_second": 21.325, "eval_steps_per_second": 0.169, "step": 20 }, { "epoch": 0.12680115273775217, "grad_norm": 0.4540252089500427, "learning_rate": 2e-05, "loss": 0.5573, "step": 22 }, { "epoch": 0.138328530259366, "grad_norm": 0.31643402576446533, "learning_rate": 2.5e-05, "loss": 0.5615, "step": 24 }, { "epoch": 0.14985590778097982, "grad_norm": 0.2575995922088623, "learning_rate": 3e-05, "loss": 0.5225, "step": 26 }, { "epoch": 0.16138328530259366, "grad_norm": 0.1953422576189041, "learning_rate": 3.5e-05, "loss": 0.5141, "step": 28 }, { "epoch": 0.1729106628242075, "grad_norm": 0.17974254488945007, "learning_rate": 4e-05, "loss": 0.508, "step": 30 }, { "epoch": 0.1844380403458213, "grad_norm": 0.16816851496696472, "learning_rate": 4.5e-05, "loss": 0.4763, "step": 32 }, { "epoch": 0.19596541786743515, "grad_norm": 0.16267944872379303, "learning_rate": 5e-05, "loss": 0.4757, "step": 34 }, { "epoch": 0.207492795389049, "grad_norm": 0.1356152594089508, "learning_rate": 4.9995821084255035e-05, "loss": 0.4714, "step": 36 }, { "epoch": 0.21902017291066284, "grad_norm": 0.12293241173028946, "learning_rate": 4.998328588931674e-05, "loss": 0.4777, "step": 38 }, { "epoch": 0.23054755043227665, "grad_norm": 0.12494528293609619, "learning_rate": 4.996239907149833e-05, "loss": 0.4649, "step": 40 }, { "epoch": 0.23054755043227665, "eval_loss": 0.45945191383361816, "eval_runtime": 163.1604, "eval_samples_per_second": 28.641, "eval_steps_per_second": 0.227, "step": 40 }, { "epoch": 0.2420749279538905, "grad_norm": 0.11425463855266571, "learning_rate": 4.993316838939995e-05, "loss": 0.4508, "step": 42 }, { "epoch": 0.25360230547550433, "grad_norm": 0.11498431116342545, "learning_rate": 4.9895604701026735e-05, "loss": 0.4433, "step": 44 }, { "epoch": 0.26512968299711814, "grad_norm": 0.12802694737911224, "learning_rate": 4.9849721959755464e-05, "loss": 0.451, "step": 46 }, { "epoch": 0.276657060518732, "grad_norm": 0.11723160743713379, "learning_rate": 4.979553720915145e-05, "loss": 0.4344, "step": 48 }, { "epoch": 0.2881844380403458, "grad_norm": 0.11519418656826019, "learning_rate": 4.9733070576637606e-05, "loss": 0.4464, "step": 50 }, { "epoch": 0.29971181556195964, "grad_norm": 0.1040327250957489, "learning_rate": 4.9662345266017836e-05, "loss": 0.4367, "step": 52 }, { "epoch": 0.3112391930835735, "grad_norm": 0.10918751358985901, "learning_rate": 4.9583387548857836e-05, "loss": 0.4411, "step": 54 }, { "epoch": 0.3227665706051873, "grad_norm": 0.11520207673311234, "learning_rate": 4.949622675472627e-05, "loss": 0.4119, "step": 56 }, { "epoch": 0.33429394812680113, "grad_norm": 0.11275804787874222, "learning_rate": 4.940089526030003e-05, "loss": 0.4387, "step": 58 }, { "epoch": 0.345821325648415, "grad_norm": 0.11030508577823639, "learning_rate": 4.92974284773376e-05, "loss": 0.4265, "step": 60 }, { "epoch": 0.345821325648415, "eval_loss": 0.4273378252983093, "eval_runtime": 163.6226, "eval_samples_per_second": 28.56, "eval_steps_per_second": 0.226, "step": 60 }, { "epoch": 0.3573487031700288, "grad_norm": 0.1050853505730629, "learning_rate": 4.9185864839525065e-05, "loss": 0.4217, "step": 62 }, { "epoch": 0.3688760806916426, "grad_norm": 0.1106187105178833, "learning_rate": 4.906624578819954e-05, "loss": 0.4244, "step": 64 }, { "epoch": 0.3804034582132565, "grad_norm": 0.1203250139951706, "learning_rate": 4.893861575695544e-05, "loss": 0.4161, "step": 66 }, { "epoch": 0.3919308357348703, "grad_norm": 0.13240748643875122, "learning_rate": 4.880302215513915e-05, "loss": 0.4205, "step": 68 }, { "epoch": 0.4034582132564842, "grad_norm": 0.11961999535560608, "learning_rate": 4.865951535023847e-05, "loss": 0.4152, "step": 70 }, { "epoch": 0.414985590778098, "grad_norm": 0.10617737472057343, "learning_rate": 4.850814864917309e-05, "loss": 0.4151, "step": 72 }, { "epoch": 0.4265129682997118, "grad_norm": 0.1053657978773117, "learning_rate": 4.834897827849325e-05, "loss": 0.4286, "step": 74 }, { "epoch": 0.43804034582132567, "grad_norm": 0.1263289898633957, "learning_rate": 4.818206336349391e-05, "loss": 0.4206, "step": 76 }, { "epoch": 0.4495677233429395, "grad_norm": 0.11088111996650696, "learning_rate": 4.8007465906252065e-05, "loss": 0.4127, "step": 78 }, { "epoch": 0.4610951008645533, "grad_norm": 0.10416624695062637, "learning_rate": 4.782525076259556e-05, "loss": 0.3924, "step": 80 }, { "epoch": 0.4610951008645533, "eval_loss": 0.41243475675582886, "eval_runtime": 163.0434, "eval_samples_per_second": 28.661, "eval_steps_per_second": 0.227, "step": 80 }, { "epoch": 0.47262247838616717, "grad_norm": 0.109275221824646, "learning_rate": 4.763548561801175e-05, "loss": 0.418, "step": 82 }, { "epoch": 0.484149855907781, "grad_norm": 0.11440867185592651, "learning_rate": 4.743824096250513e-05, "loss": 0.4054, "step": 84 }, { "epoch": 0.4956772334293948, "grad_norm": 0.11359245330095291, "learning_rate": 4.723359006441316e-05, "loss": 0.3969, "step": 86 }, { "epoch": 0.5072046109510087, "grad_norm": 0.11708330363035202, "learning_rate": 4.7021608943190105e-05, "loss": 0.4052, "step": 88 }, { "epoch": 0.5187319884726225, "grad_norm": 0.1086309477686882, "learning_rate": 4.680237634116884e-05, "loss": 0.4055, "step": 90 }, { "epoch": 0.5302593659942363, "grad_norm": 0.1133558601140976, "learning_rate": 4.6575973694311364e-05, "loss": 0.4016, "step": 92 }, { "epoch": 0.5417867435158501, "grad_norm": 0.11118675768375397, "learning_rate": 4.6342485101958634e-05, "loss": 0.4137, "step": 94 }, { "epoch": 0.553314121037464, "grad_norm": 0.1130242571234703, "learning_rate": 4.610199729559106e-05, "loss": 0.4125, "step": 96 }, { "epoch": 0.5648414985590778, "grad_norm": 0.10623504966497421, "learning_rate": 4.585459960661137e-05, "loss": 0.3989, "step": 98 }, { "epoch": 0.5763688760806917, "grad_norm": 0.10171157866716385, "learning_rate": 4.5600383933161586e-05, "loss": 0.4099, "step": 100 }, { "epoch": 0.5763688760806917, "eval_loss": 0.40298977494239807, "eval_runtime": 163.0144, "eval_samples_per_second": 28.666, "eval_steps_per_second": 0.227, "step": 100 }, { "epoch": 0.5878962536023055, "grad_norm": 0.1252756267786026, "learning_rate": 4.5339444705986636e-05, "loss": 0.4047, "step": 102 }, { "epoch": 0.5994236311239193, "grad_norm": 0.12425834685564041, "learning_rate": 4.507187885335716e-05, "loss": 0.4044, "step": 104 }, { "epoch": 0.6109510086455331, "grad_norm": 0.12188662588596344, "learning_rate": 4.479778576506467e-05, "loss": 0.3985, "step": 106 }, { "epoch": 0.622478386167147, "grad_norm": 0.1297188252210617, "learning_rate": 4.451726725550226e-05, "loss": 0.4105, "step": 108 }, { "epoch": 0.6340057636887608, "grad_norm": 0.1405544877052307, "learning_rate": 4.423042752584471e-05, "loss": 0.3943, "step": 110 }, { "epoch": 0.6455331412103746, "grad_norm": 0.1427697390317917, "learning_rate": 4.393737312534203e-05, "loss": 0.3929, "step": 112 }, { "epoch": 0.6570605187319885, "grad_norm": 0.12234554439783096, "learning_rate": 4.363821291174073e-05, "loss": 0.385, "step": 114 }, { "epoch": 0.6685878962536023, "grad_norm": 0.11913135647773743, "learning_rate": 4.3333058010847665e-05, "loss": 0.3944, "step": 116 }, { "epoch": 0.6801152737752162, "grad_norm": 0.10236409306526184, "learning_rate": 4.302202177525126e-05, "loss": 0.4088, "step": 118 }, { "epoch": 0.69164265129683, "grad_norm": 0.1148187592625618, "learning_rate": 4.270521974221572e-05, "loss": 0.3925, "step": 120 }, { "epoch": 0.69164265129683, "eval_loss": 0.39593446254730225, "eval_runtime": 163.197, "eval_samples_per_second": 28.634, "eval_steps_per_second": 0.227, "step": 120 }, { "epoch": 0.7031700288184438, "grad_norm": 0.1020105630159378, "learning_rate": 4.238276959076362e-05, "loss": 0.3959, "step": 122 }, { "epoch": 0.7146974063400576, "grad_norm": 0.09900052100419998, "learning_rate": 4.205479109796293e-05, "loss": 0.3938, "step": 124 }, { "epoch": 0.7262247838616714, "grad_norm": 0.11457400768995285, "learning_rate": 4.172140609443478e-05, "loss": 0.3941, "step": 126 }, { "epoch": 0.7377521613832853, "grad_norm": 0.12308034300804138, "learning_rate": 4.138273841909831e-05, "loss": 0.3773, "step": 128 }, { "epoch": 0.7492795389048992, "grad_norm": 0.11765991151332855, "learning_rate": 4.1038913873169606e-05, "loss": 0.3908, "step": 130 }, { "epoch": 0.760806916426513, "grad_norm": 0.11604655534029007, "learning_rate": 4.069006017343161e-05, "loss": 0.4024, "step": 132 }, { "epoch": 0.7723342939481268, "grad_norm": 0.10507753491401672, "learning_rate": 4.033630690479266e-05, "loss": 0.3874, "step": 134 }, { "epoch": 0.7838616714697406, "grad_norm": 0.11194345355033875, "learning_rate": 3.99777854721509e-05, "loss": 0.3902, "step": 136 }, { "epoch": 0.7953890489913544, "grad_norm": 0.10498375445604324, "learning_rate": 3.961462905158275e-05, "loss": 0.3922, "step": 138 }, { "epoch": 0.8069164265129684, "grad_norm": 0.10028143227100372, "learning_rate": 3.924697254087344e-05, "loss": 0.3884, "step": 140 }, { "epoch": 0.8069164265129684, "eval_loss": 0.3904818892478943, "eval_runtime": 165.7335, "eval_samples_per_second": 28.196, "eval_steps_per_second": 0.223, "step": 140 }, { "epoch": 0.8184438040345822, "grad_norm": 0.10579356551170349, "learning_rate": 3.887495250940798e-05, "loss": 0.3945, "step": 142 }, { "epoch": 0.829971181556196, "grad_norm": 0.10865680873394012, "learning_rate": 3.8498707147441186e-05, "loss": 0.396, "step": 144 }, { "epoch": 0.8414985590778098, "grad_norm": 0.10962694138288498, "learning_rate": 3.8118376214765724e-05, "loss": 0.3925, "step": 146 }, { "epoch": 0.8530259365994236, "grad_norm": 0.14297883212566376, "learning_rate": 3.773410098879694e-05, "loss": 0.3862, "step": 148 }, { "epoch": 0.8645533141210374, "grad_norm": 0.12033689022064209, "learning_rate": 3.734602421209414e-05, "loss": 0.3923, "step": 150 }, { "epoch": 0.8760806916426513, "grad_norm": 0.11992188543081284, "learning_rate": 3.6954290039337544e-05, "loss": 0.3864, "step": 152 }, { "epoch": 0.8876080691642652, "grad_norm": 0.1202910915017128, "learning_rate": 3.6559043983780695e-05, "loss": 0.3965, "step": 154 }, { "epoch": 0.899135446685879, "grad_norm": 0.1139882281422615, "learning_rate": 3.6160432863198227e-05, "loss": 0.3867, "step": 156 }, { "epoch": 0.9106628242074928, "grad_norm": 0.10837512463331223, "learning_rate": 3.575860474534907e-05, "loss": 0.3734, "step": 158 }, { "epoch": 0.9221902017291066, "grad_norm": 0.11608846485614777, "learning_rate": 3.535370889297532e-05, "loss": 0.3856, "step": 160 }, { "epoch": 0.9221902017291066, "eval_loss": 0.3864249289035797, "eval_runtime": 162.9187, "eval_samples_per_second": 28.683, "eval_steps_per_second": 0.227, "step": 160 }, { "epoch": 0.9337175792507204, "grad_norm": 0.11061898618936539, "learning_rate": 3.494589570835719e-05, "loss": 0.3874, "step": 162 }, { "epoch": 0.9452449567723343, "grad_norm": 0.10494677722454071, "learning_rate": 3.4535316677444745e-05, "loss": 0.3862, "step": 164 }, { "epoch": 0.9567723342939481, "grad_norm": 0.11414439976215363, "learning_rate": 3.412212431358704e-05, "loss": 0.382, "step": 166 }, { "epoch": 0.968299711815562, "grad_norm": 0.12404631823301315, "learning_rate": 3.3706472100879635e-05, "loss": 0.3893, "step": 168 }, { "epoch": 0.9798270893371758, "grad_norm": 0.11317316442728043, "learning_rate": 3.3288514437151505e-05, "loss": 0.379, "step": 170 }, { "epoch": 0.9913544668587896, "grad_norm": 0.11200971901416779, "learning_rate": 3.286840657661259e-05, "loss": 0.3829, "step": 172 }, { "epoch": 1.0028818443804035, "grad_norm": 0.11575999855995178, "learning_rate": 3.2446304572183155e-05, "loss": 0.3757, "step": 174 }, { "epoch": 1.0144092219020173, "grad_norm": 0.1118895411491394, "learning_rate": 3.2022365217526515e-05, "loss": 0.3487, "step": 176 }, { "epoch": 1.0259365994236311, "grad_norm": 0.11941341310739517, "learning_rate": 3.159674598880658e-05, "loss": 0.3611, "step": 178 }, { "epoch": 1.037463976945245, "grad_norm": 0.12434028834104538, "learning_rate": 3.116960498619191e-05, "loss": 0.3527, "step": 180 }, { "epoch": 1.037463976945245, "eval_loss": 0.38433459401130676, "eval_runtime": 167.462, "eval_samples_per_second": 27.905, "eval_steps_per_second": 0.221, "step": 180 }, { "epoch": 1.0489913544668588, "grad_norm": 0.1257391721010208, "learning_rate": 3.0741100875127956e-05, "loss": 0.346, "step": 182 }, { "epoch": 1.0605187319884726, "grad_norm": 0.1072888970375061, "learning_rate": 3.0311392827399266e-05, "loss": 0.351, "step": 184 }, { "epoch": 1.0720461095100864, "grad_norm": 0.11550486832857132, "learning_rate": 2.9880640462003766e-05, "loss": 0.3584, "step": 186 }, { "epoch": 1.0835734870317002, "grad_norm": 0.11017844080924988, "learning_rate": 2.944900378586073e-05, "loss": 0.3528, "step": 188 }, { "epoch": 1.0951008645533142, "grad_norm": 0.11351864784955978, "learning_rate": 2.901664313437478e-05, "loss": 0.3548, "step": 190 }, { "epoch": 1.106628242074928, "grad_norm": 0.11179786920547485, "learning_rate": 2.8583719111877844e-05, "loss": 0.3433, "step": 192 }, { "epoch": 1.1181556195965419, "grad_norm": 0.10140883177518845, "learning_rate": 2.8150392531971137e-05, "loss": 0.3524, "step": 194 }, { "epoch": 1.1296829971181557, "grad_norm": 0.10281901806592941, "learning_rate": 2.7716824357789567e-05, "loss": 0.3527, "step": 196 }, { "epoch": 1.1412103746397695, "grad_norm": 0.09435153007507324, "learning_rate": 2.728317564221044e-05, "loss": 0.3521, "step": 198 }, { "epoch": 1.1527377521613833, "grad_norm": 0.10692735016345978, "learning_rate": 2.684960746802887e-05, "loss": 0.3635, "step": 200 }, { "epoch": 1.1527377521613833, "eval_loss": 0.3818141520023346, "eval_runtime": 163.2711, "eval_samples_per_second": 28.621, "eval_steps_per_second": 0.227, "step": 200 }, { "epoch": 1.1642651296829971, "grad_norm": 0.10303134471178055, "learning_rate": 2.6416280888122165e-05, "loss": 0.341, "step": 202 }, { "epoch": 1.175792507204611, "grad_norm": 0.1047578901052475, "learning_rate": 2.5983356865625224e-05, "loss": 0.3483, "step": 204 }, { "epoch": 1.1873198847262247, "grad_norm": 0.09823092073202133, "learning_rate": 2.5550996214139283e-05, "loss": 0.3469, "step": 206 }, { "epoch": 1.1988472622478386, "grad_norm": 0.10456107556819916, "learning_rate": 2.511935953799624e-05, "loss": 0.3498, "step": 208 }, { "epoch": 1.2103746397694524, "grad_norm": 0.10154879838228226, "learning_rate": 2.4688607172600742e-05, "loss": 0.3494, "step": 210 }, { "epoch": 1.2219020172910664, "grad_norm": 0.10301995277404785, "learning_rate": 2.4258899124872063e-05, "loss": 0.344, "step": 212 }, { "epoch": 1.23342939481268, "grad_norm": 0.10198619216680527, "learning_rate": 2.3830395013808088e-05, "loss": 0.3453, "step": 214 }, { "epoch": 1.244956772334294, "grad_norm": 0.10618147999048233, "learning_rate": 2.3403254011193433e-05, "loss": 0.3492, "step": 216 }, { "epoch": 1.2564841498559078, "grad_norm": 0.10644455999135971, "learning_rate": 2.2977634782473493e-05, "loss": 0.3365, "step": 218 }, { "epoch": 1.2680115273775217, "grad_norm": 0.10750509053468704, "learning_rate": 2.2553695427816847e-05, "loss": 0.3527, "step": 220 }, { "epoch": 1.2680115273775217, "eval_loss": 0.37958407402038574, "eval_runtime": 162.0488, "eval_samples_per_second": 28.837, "eval_steps_per_second": 0.228, "step": 220 }, { "epoch": 1.2795389048991355, "grad_norm": 0.10969573259353638, "learning_rate": 2.2131593423387414e-05, "loss": 0.3388, "step": 222 }, { "epoch": 1.2910662824207493, "grad_norm": 0.11101813614368439, "learning_rate": 2.1711485562848504e-05, "loss": 0.3417, "step": 224 }, { "epoch": 1.302593659942363, "grad_norm": 0.10727708041667938, "learning_rate": 2.1293527899120374e-05, "loss": 0.3478, "step": 226 }, { "epoch": 1.314121037463977, "grad_norm": 0.11611964553594589, "learning_rate": 2.0877875686412966e-05, "loss": 0.34, "step": 228 }, { "epoch": 1.3256484149855907, "grad_norm": 0.11851979792118073, "learning_rate": 2.0464683322555267e-05, "loss": 0.3413, "step": 230 }, { "epoch": 1.3371757925072045, "grad_norm": 0.11880774050951004, "learning_rate": 2.0054104291642818e-05, "loss": 0.3479, "step": 232 }, { "epoch": 1.3487031700288186, "grad_norm": 0.10287457704544067, "learning_rate": 1.964629110702469e-05, "loss": 0.3434, "step": 234 }, { "epoch": 1.3602305475504322, "grad_norm": 0.10474801808595657, "learning_rate": 1.924139525465093e-05, "loss": 0.338, "step": 236 }, { "epoch": 1.3717579250720462, "grad_norm": 0.10204192996025085, "learning_rate": 1.8839567136801772e-05, "loss": 0.3369, "step": 238 }, { "epoch": 1.38328530259366, "grad_norm": 0.1050105094909668, "learning_rate": 1.8440956016219314e-05, "loss": 0.3407, "step": 240 }, { "epoch": 1.38328530259366, "eval_loss": 0.3772360384464264, "eval_runtime": 161.7846, "eval_samples_per_second": 28.884, "eval_steps_per_second": 0.229, "step": 240 }, { "epoch": 1.3948126801152738, "grad_norm": 0.10698170959949493, "learning_rate": 1.8045709960662465e-05, "loss": 0.355, "step": 242 }, { "epoch": 1.4063400576368876, "grad_norm": 0.09979532659053802, "learning_rate": 1.765397578790587e-05, "loss": 0.3419, "step": 244 }, { "epoch": 1.4178674351585014, "grad_norm": 0.09883076697587967, "learning_rate": 1.7265899011203065e-05, "loss": 0.3415, "step": 246 }, { "epoch": 1.4293948126801153, "grad_norm": 0.10558874905109406, "learning_rate": 1.6881623785234285e-05, "loss": 0.353, "step": 248 }, { "epoch": 1.440922190201729, "grad_norm": 0.1028595045208931, "learning_rate": 1.650129285255882e-05, "loss": 0.3537, "step": 250 }, { "epoch": 1.4524495677233429, "grad_norm": 0.10255461186170578, "learning_rate": 1.6125047490592034e-05, "loss": 0.3393, "step": 252 }, { "epoch": 1.4639769452449567, "grad_norm": 0.10520190000534058, "learning_rate": 1.5753027459126566e-05, "loss": 0.3469, "step": 254 }, { "epoch": 1.4755043227665707, "grad_norm": 0.10457232594490051, "learning_rate": 1.5385370948417256e-05, "loss": 0.3548, "step": 256 }, { "epoch": 1.4870317002881843, "grad_norm": 0.10024940967559814, "learning_rate": 1.5022214527849105e-05, "loss": 0.3403, "step": 258 }, { "epoch": 1.4985590778097984, "grad_norm": 0.09729419648647308, "learning_rate": 1.4663693095207343e-05, "loss": 0.3372, "step": 260 }, { "epoch": 1.4985590778097984, "eval_loss": 0.3750576674938202, "eval_runtime": 161.7883, "eval_samples_per_second": 28.883, "eval_steps_per_second": 0.229, "step": 260 }, { "epoch": 1.510086455331412, "grad_norm": 0.09922486543655396, "learning_rate": 1.43099398265684e-05, "loss": 0.329, "step": 262 }, { "epoch": 1.521613832853026, "grad_norm": 0.10505367070436478, "learning_rate": 1.3961086126830403e-05, "loss": 0.3404, "step": 264 }, { "epoch": 1.5331412103746398, "grad_norm": 0.10513550788164139, "learning_rate": 1.3617261580901691e-05, "loss": 0.3436, "step": 266 }, { "epoch": 1.5446685878962536, "grad_norm": 0.10580945760011673, "learning_rate": 1.3278593905565224e-05, "loss": 0.3448, "step": 268 }, { "epoch": 1.5561959654178674, "grad_norm": 0.0965629443526268, "learning_rate": 1.2945208902037071e-05, "loss": 0.3323, "step": 270 }, { "epoch": 1.5677233429394812, "grad_norm": 0.1028643250465393, "learning_rate": 1.261723040923638e-05, "loss": 0.341, "step": 272 }, { "epoch": 1.579250720461095, "grad_norm": 0.10219055414199829, "learning_rate": 1.2294780257784277e-05, "loss": 0.3337, "step": 274 }, { "epoch": 1.5907780979827089, "grad_norm": 0.10359715670347214, "learning_rate": 1.1977978224748735e-05, "loss": 0.3388, "step": 276 }, { "epoch": 1.602305475504323, "grad_norm": 0.09491585195064545, "learning_rate": 1.166694198915234e-05, "loss": 0.3404, "step": 278 }, { "epoch": 1.6138328530259365, "grad_norm": 0.10194993764162064, "learning_rate": 1.1361787088259273e-05, "loss": 0.3453, "step": 280 }, { "epoch": 1.6138328530259365, "eval_loss": 0.3727053105831146, "eval_runtime": 161.5061, "eval_samples_per_second": 28.934, "eval_steps_per_second": 0.229, "step": 280 }, { "epoch": 1.6253602305475505, "grad_norm": 0.09746123105287552, "learning_rate": 1.1062626874657977e-05, "loss": 0.3372, "step": 282 }, { "epoch": 1.6368876080691641, "grad_norm": 0.1048843190073967, "learning_rate": 1.0769572474155296e-05, "loss": 0.3415, "step": 284 }, { "epoch": 1.6484149855907781, "grad_norm": 0.10008594393730164, "learning_rate": 1.0482732744497742e-05, "loss": 0.3332, "step": 286 }, { "epoch": 1.659942363112392, "grad_norm": 0.09605535864830017, "learning_rate": 1.020221423493533e-05, "loss": 0.3355, "step": 288 }, { "epoch": 1.6714697406340058, "grad_norm": 0.0957765281200409, "learning_rate": 9.928121146642841e-06, "loss": 0.3511, "step": 290 }, { "epoch": 1.6829971181556196, "grad_norm": 0.10064635425806046, "learning_rate": 9.660555294013373e-06, "loss": 0.3462, "step": 292 }, { "epoch": 1.6945244956772334, "grad_norm": 0.09759914875030518, "learning_rate": 9.399616066838415e-06, "loss": 0.3476, "step": 294 }, { "epoch": 1.7060518731988472, "grad_norm": 0.1018439456820488, "learning_rate": 9.145400393388629e-06, "loss": 0.3542, "step": 296 }, { "epoch": 1.717579250720461, "grad_norm": 0.10098982602357864, "learning_rate": 8.898002704408944e-06, "loss": 0.3474, "step": 298 }, { "epoch": 1.729106628242075, "grad_norm": 0.09507515281438828, "learning_rate": 8.657514898041381e-06, "loss": 0.3408, "step": 300 }, { "epoch": 1.729106628242075, "eval_loss": 0.37131258845329285, "eval_runtime": 161.848, "eval_samples_per_second": 28.873, "eval_steps_per_second": 0.229, "step": 300 }, { "epoch": 1.7406340057636887, "grad_norm": 0.09314768761396408, "learning_rate": 8.424026305688643e-06, "loss": 0.341, "step": 302 }, { "epoch": 1.7521613832853027, "grad_norm": 0.09930098056793213, "learning_rate": 8.197623658831165e-06, "loss": 0.3397, "step": 304 }, { "epoch": 1.7636887608069163, "grad_norm": 0.10165336728096008, "learning_rate": 7.978391056809904e-06, "loss": 0.3419, "step": 306 }, { "epoch": 1.7752161383285303, "grad_norm": 0.09817034006118774, "learning_rate": 7.766409935586837e-06, "loss": 0.3456, "step": 308 }, { "epoch": 1.7867435158501441, "grad_norm": 0.10100841522216797, "learning_rate": 7.5617590374948745e-06, "loss": 0.3464, "step": 310 }, { "epoch": 1.798270893371758, "grad_norm": 0.10015368461608887, "learning_rate": 7.364514381988254e-06, "loss": 0.3402, "step": 312 }, { "epoch": 1.8097982708933718, "grad_norm": 0.10384070128202438, "learning_rate": 7.174749237404444e-06, "loss": 0.3312, "step": 314 }, { "epoch": 1.8213256484149856, "grad_norm": 0.10439052432775497, "learning_rate": 6.992534093747942e-06, "loss": 0.3333, "step": 316 }, { "epoch": 1.8328530259365994, "grad_norm": 0.09687253832817078, "learning_rate": 6.8179366365061e-06, "loss": 0.3358, "step": 318 }, { "epoch": 1.8443804034582132, "grad_norm": 0.10308189690113068, "learning_rate": 6.651021721506756e-06, "loss": 0.3342, "step": 320 }, { "epoch": 1.8443804034582132, "eval_loss": 0.3700437545776367, "eval_runtime": 161.117, "eval_samples_per_second": 29.004, "eval_steps_per_second": 0.23, "step": 320 }, { "epoch": 1.8559077809798272, "grad_norm": 0.09917671233415604, "learning_rate": 6.491851350826915e-06, "loss": 0.35, "step": 322 }, { "epoch": 1.8674351585014408, "grad_norm": 0.09794910252094269, "learning_rate": 6.340484649761529e-06, "loss": 0.3382, "step": 324 }, { "epoch": 1.8789625360230549, "grad_norm": 0.10062103718519211, "learning_rate": 6.196977844860848e-06, "loss": 0.3313, "step": 326 }, { "epoch": 1.8904899135446684, "grad_norm": 0.09653456509113312, "learning_rate": 6.061384243044564e-06, "loss": 0.3364, "step": 328 }, { "epoch": 1.9020172910662825, "grad_norm": 0.09764689952135086, "learning_rate": 5.933754211800459e-06, "loss": 0.332, "step": 330 }, { "epoch": 1.9135446685878963, "grad_norm": 0.09577332437038422, "learning_rate": 5.814135160474937e-06, "loss": 0.339, "step": 332 }, { "epoch": 1.92507204610951, "grad_norm": 0.0969650149345398, "learning_rate": 5.7025715226624036e-06, "loss": 0.3301, "step": 334 }, { "epoch": 1.936599423631124, "grad_norm": 0.09393922984600067, "learning_rate": 5.5991047396999735e-06, "loss": 0.3374, "step": 336 }, { "epoch": 1.9481268011527377, "grad_norm": 0.11683030426502228, "learning_rate": 5.503773245273734e-06, "loss": 0.3424, "step": 338 }, { "epoch": 1.9596541786743515, "grad_norm": 0.09428226202726364, "learning_rate": 5.41661245114217e-06, "loss": 0.332, "step": 340 }, { "epoch": 1.9596541786743515, "eval_loss": 0.36898180842399597, "eval_runtime": 161.0051, "eval_samples_per_second": 29.024, "eval_steps_per_second": 0.23, "step": 340 }, { "epoch": 1.9711815561959654, "grad_norm": 0.09828232228755951, "learning_rate": 5.337654733982173e-06, "loss": 0.3386, "step": 342 }, { "epoch": 1.9827089337175794, "grad_norm": 0.09727545082569122, "learning_rate": 5.2669294233624e-06, "loss": 0.331, "step": 344 }, { "epoch": 1.994236311239193, "grad_norm": 0.09713173657655716, "learning_rate": 5.204462790848549e-06, "loss": 0.3478, "step": 346 }, { "epoch": 1.994236311239193, "step": 346, "total_flos": 388981471576064.0, "train_loss": 0.3994986457976303, "train_runtime": 42730.5756, "train_samples_per_second": 4.155, "train_steps_per_second": 0.008 } ], "logging_steps": 2, "max_steps": 346, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 388981471576064.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }