{ "best_metric": 0.9384466409683228, "best_model_checkpoint": "/kaggle/working/LLaMA-Factory/output/dolphin-2.9-llama3-8b-GER/checkpoint-800", "epoch": 1.0, "eval_steps": 100, "global_step": 849, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 0.5445218086242676, "learning_rate": 2.5e-06, "loss": 1.2105, "step": 1 }, { "epoch": 0.0, "grad_norm": 0.4362960457801819, "learning_rate": 5e-06, "loss": 0.9311, "step": 2 }, { "epoch": 0.0, "grad_norm": 0.5098525881767273, "learning_rate": 7.5e-06, "loss": 1.3604, "step": 3 }, { "epoch": 0.0, "grad_norm": 1.0830402374267578, "learning_rate": 1e-05, "loss": 1.5808, "step": 4 }, { "epoch": 0.01, "grad_norm": 0.5936903953552246, "learning_rate": 1.25e-05, "loss": 1.248, "step": 5 }, { "epoch": 0.01, "grad_norm": 0.7212775945663452, "learning_rate": 1.5e-05, "loss": 1.2712, "step": 6 }, { "epoch": 0.01, "grad_norm": 0.4733128845691681, "learning_rate": 1.75e-05, "loss": 1.3005, "step": 7 }, { "epoch": 0.01, "grad_norm": 0.6236618757247925, "learning_rate": 2e-05, "loss": 1.3398, "step": 8 }, { "epoch": 0.01, "grad_norm": 0.5326029658317566, "learning_rate": 2.25e-05, "loss": 1.0476, "step": 9 }, { "epoch": 0.01, "grad_norm": 0.5021428465843201, "learning_rate": 2.5e-05, "loss": 1.1245, "step": 10 }, { "epoch": 0.01, "grad_norm": 0.46588796377182007, "learning_rate": 2.7500000000000004e-05, "loss": 1.1041, "step": 11 }, { "epoch": 0.01, "grad_norm": 0.936045229434967, "learning_rate": 3e-05, "loss": 1.3201, "step": 12 }, { "epoch": 0.02, "grad_norm": 0.5649005770683289, "learning_rate": 3.2500000000000004e-05, "loss": 1.1596, "step": 13 }, { "epoch": 0.02, "grad_norm": 0.5679148435592651, "learning_rate": 3.5e-05, "loss": 1.2501, "step": 14 }, { "epoch": 0.02, "grad_norm": 0.8638319373130798, "learning_rate": 3.7500000000000003e-05, "loss": 1.237, "step": 15 }, { "epoch": 0.02, "grad_norm": 0.5722486972808838, "learning_rate": 4e-05, "loss": 1.2616, "step": 16 }, { "epoch": 0.02, "grad_norm": 0.5802999138832092, "learning_rate": 4.25e-05, "loss": 0.9578, "step": 17 }, { "epoch": 0.02, "grad_norm": 1.2806150913238525, "learning_rate": 4.5e-05, "loss": 1.3815, "step": 18 }, { "epoch": 0.02, "grad_norm": 0.5839210748672485, "learning_rate": 4.75e-05, "loss": 1.2619, "step": 19 }, { "epoch": 0.02, "grad_norm": 0.7132797837257385, "learning_rate": 5e-05, "loss": 1.2842, "step": 20 }, { "epoch": 0.02, "grad_norm": 0.7936571836471558, "learning_rate": 5.25e-05, "loss": 1.3991, "step": 21 }, { "epoch": 0.03, "grad_norm": 0.7268731594085693, "learning_rate": 5.500000000000001e-05, "loss": 1.2217, "step": 22 }, { "epoch": 0.03, "grad_norm": 0.6649657487869263, "learning_rate": 5.7499999999999995e-05, "loss": 1.2485, "step": 23 }, { "epoch": 0.03, "grad_norm": 0.45912331342697144, "learning_rate": 6e-05, "loss": 1.1452, "step": 24 }, { "epoch": 0.03, "grad_norm": 0.49475613236427307, "learning_rate": 6.25e-05, "loss": 1.256, "step": 25 }, { "epoch": 0.03, "grad_norm": 0.6027999520301819, "learning_rate": 6.500000000000001e-05, "loss": 1.2267, "step": 26 }, { "epoch": 0.03, "grad_norm": 0.48380520939826965, "learning_rate": 6.750000000000001e-05, "loss": 1.202, "step": 27 }, { "epoch": 0.03, "grad_norm": 0.6623361706733704, "learning_rate": 7e-05, "loss": 1.3924, "step": 28 }, { "epoch": 0.03, "grad_norm": 0.6632773280143738, "learning_rate": 7.25e-05, "loss": 1.2443, "step": 29 }, { "epoch": 0.04, "grad_norm": 0.5974243879318237, "learning_rate": 7.500000000000001e-05, "loss": 1.0591, "step": 30 }, { "epoch": 0.04, "grad_norm": 0.7902927398681641, "learning_rate": 7.75e-05, "loss": 1.0192, "step": 31 }, { "epoch": 0.04, "grad_norm": 0.7081462144851685, "learning_rate": 8e-05, "loss": 1.0955, "step": 32 }, { "epoch": 0.04, "grad_norm": 0.4744333326816559, "learning_rate": 8.25e-05, "loss": 1.0564, "step": 33 }, { "epoch": 0.04, "grad_norm": 0.703999400138855, "learning_rate": 8.5e-05, "loss": 1.1022, "step": 34 }, { "epoch": 0.04, "grad_norm": 0.5378603339195251, "learning_rate": 8.75e-05, "loss": 1.063, "step": 35 }, { "epoch": 0.04, "grad_norm": 0.6105419397354126, "learning_rate": 9e-05, "loss": 0.9748, "step": 36 }, { "epoch": 0.04, "grad_norm": 0.6250177025794983, "learning_rate": 9.250000000000001e-05, "loss": 0.9951, "step": 37 }, { "epoch": 0.04, "grad_norm": 0.6306005716323853, "learning_rate": 9.5e-05, "loss": 0.9875, "step": 38 }, { "epoch": 0.05, "grad_norm": 0.8635210990905762, "learning_rate": 9.75e-05, "loss": 1.1825, "step": 39 }, { "epoch": 0.05, "grad_norm": 0.5938963890075684, "learning_rate": 0.0001, "loss": 0.9678, "step": 40 }, { "epoch": 0.05, "grad_norm": 0.5765062570571899, "learning_rate": 0.0001025, "loss": 1.1031, "step": 41 }, { "epoch": 0.05, "grad_norm": 0.5162666440010071, "learning_rate": 0.000105, "loss": 1.0831, "step": 42 }, { "epoch": 0.05, "grad_norm": 0.8675170540809631, "learning_rate": 0.0001075, "loss": 1.0034, "step": 43 }, { "epoch": 0.05, "grad_norm": 0.44026264548301697, "learning_rate": 0.00011000000000000002, "loss": 0.9607, "step": 44 }, { "epoch": 0.05, "grad_norm": 0.4922897219657898, "learning_rate": 0.00011250000000000001, "loss": 1.0251, "step": 45 }, { "epoch": 0.05, "grad_norm": 0.5505048632621765, "learning_rate": 0.00011499999999999999, "loss": 0.6948, "step": 46 }, { "epoch": 0.06, "grad_norm": 0.5226988792419434, "learning_rate": 0.00011750000000000001, "loss": 1.2259, "step": 47 }, { "epoch": 0.06, "grad_norm": 0.5105571150779724, "learning_rate": 0.00012, "loss": 1.1239, "step": 48 }, { "epoch": 0.06, "grad_norm": 2.0914857387542725, "learning_rate": 0.00012250000000000002, "loss": 1.4032, "step": 49 }, { "epoch": 0.06, "grad_norm": 0.4611998498439789, "learning_rate": 0.000125, "loss": 1.1481, "step": 50 }, { "epoch": 0.06, "grad_norm": 0.5243968367576599, "learning_rate": 0.0001275, "loss": 1.0514, "step": 51 }, { "epoch": 0.06, "grad_norm": 0.48045778274536133, "learning_rate": 0.00013000000000000002, "loss": 1.1552, "step": 52 }, { "epoch": 0.06, "grad_norm": 0.5543797612190247, "learning_rate": 0.0001325, "loss": 1.035, "step": 53 }, { "epoch": 0.06, "grad_norm": 0.5148913264274597, "learning_rate": 0.00013500000000000003, "loss": 0.9244, "step": 54 }, { "epoch": 0.06, "grad_norm": 0.5708130598068237, "learning_rate": 0.0001375, "loss": 1.2272, "step": 55 }, { "epoch": 0.07, "grad_norm": 0.7507888078689575, "learning_rate": 0.00014, "loss": 1.1196, "step": 56 }, { "epoch": 0.07, "grad_norm": 0.5411068201065063, "learning_rate": 0.00014250000000000002, "loss": 1.1444, "step": 57 }, { "epoch": 0.07, "grad_norm": 0.7314138412475586, "learning_rate": 0.000145, "loss": 1.3451, "step": 58 }, { "epoch": 0.07, "grad_norm": 0.5134022831916809, "learning_rate": 0.0001475, "loss": 0.8021, "step": 59 }, { "epoch": 0.07, "grad_norm": 0.7859703302383423, "learning_rate": 0.00015000000000000001, "loss": 0.952, "step": 60 }, { "epoch": 0.07, "grad_norm": 0.4740566611289978, "learning_rate": 0.0001525, "loss": 0.8878, "step": 61 }, { "epoch": 0.07, "grad_norm": 0.8245677947998047, "learning_rate": 0.000155, "loss": 1.4341, "step": 62 }, { "epoch": 0.07, "grad_norm": 0.5067218542098999, "learning_rate": 0.0001575, "loss": 0.8404, "step": 63 }, { "epoch": 0.08, "grad_norm": 0.40824806690216064, "learning_rate": 0.00016, "loss": 1.0943, "step": 64 }, { "epoch": 0.08, "grad_norm": 0.428151935338974, "learning_rate": 0.00016250000000000002, "loss": 1.0819, "step": 65 }, { "epoch": 0.08, "grad_norm": 0.437532901763916, "learning_rate": 0.000165, "loss": 1.0461, "step": 66 }, { "epoch": 0.08, "grad_norm": 0.4619809091091156, "learning_rate": 0.0001675, "loss": 1.0232, "step": 67 }, { "epoch": 0.08, "grad_norm": 1.3247042894363403, "learning_rate": 0.00017, "loss": 1.218, "step": 68 }, { "epoch": 0.08, "grad_norm": 0.5896454453468323, "learning_rate": 0.00017250000000000002, "loss": 1.1953, "step": 69 }, { "epoch": 0.08, "grad_norm": 0.5605736970901489, "learning_rate": 0.000175, "loss": 1.2542, "step": 70 }, { "epoch": 0.08, "grad_norm": 0.36923155188560486, "learning_rate": 0.0001775, "loss": 0.7096, "step": 71 }, { "epoch": 0.08, "grad_norm": 0.6809273362159729, "learning_rate": 0.00018, "loss": 1.1959, "step": 72 }, { "epoch": 0.09, "grad_norm": 0.42645493149757385, "learning_rate": 0.0001825, "loss": 1.1829, "step": 73 }, { "epoch": 0.09, "grad_norm": 0.7621450424194336, "learning_rate": 0.00018500000000000002, "loss": 1.1599, "step": 74 }, { "epoch": 0.09, "grad_norm": 0.5299301743507385, "learning_rate": 0.0001875, "loss": 1.0546, "step": 75 }, { "epoch": 0.09, "grad_norm": 0.484611451625824, "learning_rate": 0.00019, "loss": 1.0854, "step": 76 }, { "epoch": 0.09, "grad_norm": 0.4027417004108429, "learning_rate": 0.00019250000000000002, "loss": 0.7812, "step": 77 }, { "epoch": 0.09, "grad_norm": 0.7537221908569336, "learning_rate": 0.000195, "loss": 0.9719, "step": 78 }, { "epoch": 0.09, "grad_norm": 0.45811474323272705, "learning_rate": 0.00019750000000000003, "loss": 0.9684, "step": 79 }, { "epoch": 0.09, "grad_norm": 0.5426852107048035, "learning_rate": 0.0002, "loss": 0.9451, "step": 80 }, { "epoch": 0.1, "grad_norm": 0.48964977264404297, "learning_rate": 0.000199739921976593, "loss": 1.0807, "step": 81 }, { "epoch": 0.1, "grad_norm": 0.6225113272666931, "learning_rate": 0.00019947984395318597, "loss": 1.1913, "step": 82 }, { "epoch": 0.1, "grad_norm": 0.47157713770866394, "learning_rate": 0.00019921976592977893, "loss": 1.0392, "step": 83 }, { "epoch": 0.1, "grad_norm": 0.42048129439353943, "learning_rate": 0.0001989596879063719, "loss": 0.8439, "step": 84 }, { "epoch": 0.1, "grad_norm": 0.4129664897918701, "learning_rate": 0.0001986996098829649, "loss": 1.1025, "step": 85 }, { "epoch": 0.1, "grad_norm": 0.5610629320144653, "learning_rate": 0.00019843953185955788, "loss": 0.9368, "step": 86 }, { "epoch": 0.1, "grad_norm": 0.4288642406463623, "learning_rate": 0.00019817945383615086, "loss": 0.9711, "step": 87 }, { "epoch": 0.1, "grad_norm": 0.572507917881012, "learning_rate": 0.0001979193758127438, "loss": 0.9739, "step": 88 }, { "epoch": 0.1, "grad_norm": 0.5115912556648254, "learning_rate": 0.0001976592977893368, "loss": 0.9356, "step": 89 }, { "epoch": 0.11, "grad_norm": 0.608063280582428, "learning_rate": 0.00019739921976592978, "loss": 1.2459, "step": 90 }, { "epoch": 0.11, "grad_norm": 1.0171502828598022, "learning_rate": 0.00019713914174252276, "loss": 1.0629, "step": 91 }, { "epoch": 0.11, "grad_norm": 0.514552116394043, "learning_rate": 0.00019687906371911574, "loss": 0.9646, "step": 92 }, { "epoch": 0.11, "grad_norm": 0.4363428056240082, "learning_rate": 0.00019661898569570872, "loss": 0.8659, "step": 93 }, { "epoch": 0.11, "grad_norm": 0.4470008611679077, "learning_rate": 0.0001963589076723017, "loss": 1.0338, "step": 94 }, { "epoch": 0.11, "grad_norm": 0.5710871815681458, "learning_rate": 0.0001960988296488947, "loss": 0.946, "step": 95 }, { "epoch": 0.11, "grad_norm": 0.6456964612007141, "learning_rate": 0.00019583875162548767, "loss": 1.0956, "step": 96 }, { "epoch": 0.11, "grad_norm": 0.45742425322532654, "learning_rate": 0.00019557867360208065, "loss": 0.7985, "step": 97 }, { "epoch": 0.12, "grad_norm": 0.9953218698501587, "learning_rate": 0.0001953185955786736, "loss": 0.9476, "step": 98 }, { "epoch": 0.12, "grad_norm": 0.5075244307518005, "learning_rate": 0.0001950585175552666, "loss": 1.1704, "step": 99 }, { "epoch": 0.12, "grad_norm": 0.7076822519302368, "learning_rate": 0.00019479843953185957, "loss": 1.2054, "step": 100 }, { "epoch": 0.12, "eval_loss": 1.0368528366088867, "eval_runtime": 547.1826, "eval_samples_per_second": 2.76, "eval_steps_per_second": 1.38, "step": 100 }, { "epoch": 0.12, "grad_norm": 0.7408495545387268, "learning_rate": 0.00019453836150845255, "loss": 1.3149, "step": 101 }, { "epoch": 0.12, "grad_norm": 0.5010321736335754, "learning_rate": 0.00019427828348504554, "loss": 0.8483, "step": 102 }, { "epoch": 0.12, "grad_norm": 0.5388755798339844, "learning_rate": 0.0001940182054616385, "loss": 1.034, "step": 103 }, { "epoch": 0.12, "grad_norm": 0.32918357849121094, "learning_rate": 0.00019375812743823147, "loss": 1.0762, "step": 104 }, { "epoch": 0.12, "grad_norm": 0.4812282919883728, "learning_rate": 0.00019349804941482446, "loss": 1.0797, "step": 105 }, { "epoch": 0.12, "grad_norm": 0.6160157918930054, "learning_rate": 0.00019323797139141744, "loss": 0.9488, "step": 106 }, { "epoch": 0.13, "grad_norm": 0.41306182742118835, "learning_rate": 0.00019297789336801042, "loss": 1.0742, "step": 107 }, { "epoch": 0.13, "grad_norm": 0.5411396026611328, "learning_rate": 0.00019271781534460338, "loss": 1.1162, "step": 108 }, { "epoch": 0.13, "grad_norm": 0.45688003301620483, "learning_rate": 0.00019245773732119636, "loss": 0.8042, "step": 109 }, { "epoch": 0.13, "grad_norm": 0.4660446047782898, "learning_rate": 0.00019219765929778934, "loss": 0.8234, "step": 110 }, { "epoch": 0.13, "grad_norm": 0.6277762055397034, "learning_rate": 0.00019193758127438232, "loss": 1.0324, "step": 111 }, { "epoch": 0.13, "grad_norm": 0.6119320392608643, "learning_rate": 0.0001916775032509753, "loss": 0.6912, "step": 112 }, { "epoch": 0.13, "grad_norm": 0.5463618040084839, "learning_rate": 0.00019141742522756826, "loss": 1.1297, "step": 113 }, { "epoch": 0.13, "grad_norm": 0.4918029010295868, "learning_rate": 0.00019115734720416124, "loss": 1.0857, "step": 114 }, { "epoch": 0.14, "grad_norm": 0.4734187424182892, "learning_rate": 0.00019089726918075422, "loss": 1.0246, "step": 115 }, { "epoch": 0.14, "grad_norm": 0.5416173934936523, "learning_rate": 0.0001906371911573472, "loss": 1.0262, "step": 116 }, { "epoch": 0.14, "grad_norm": 0.486284464597702, "learning_rate": 0.0001903771131339402, "loss": 1.2198, "step": 117 }, { "epoch": 0.14, "grad_norm": 0.4003067910671234, "learning_rate": 0.00019011703511053317, "loss": 0.8525, "step": 118 }, { "epoch": 0.14, "grad_norm": 0.38142916560173035, "learning_rate": 0.00018985695708712615, "loss": 1.1475, "step": 119 }, { "epoch": 0.14, "grad_norm": 0.42364147305488586, "learning_rate": 0.00018959687906371913, "loss": 0.8207, "step": 120 }, { "epoch": 0.14, "grad_norm": 0.384432852268219, "learning_rate": 0.00018933680104031212, "loss": 0.9451, "step": 121 }, { "epoch": 0.14, "grad_norm": 0.38188567757606506, "learning_rate": 0.0001890767230169051, "loss": 0.921, "step": 122 }, { "epoch": 0.14, "grad_norm": 0.4212167263031006, "learning_rate": 0.00018881664499349805, "loss": 0.8532, "step": 123 }, { "epoch": 0.15, "grad_norm": 0.7674171924591064, "learning_rate": 0.00018855656697009104, "loss": 1.1751, "step": 124 }, { "epoch": 0.15, "grad_norm": 0.4111984074115753, "learning_rate": 0.00018829648894668402, "loss": 0.906, "step": 125 }, { "epoch": 0.15, "grad_norm": 0.6682639718055725, "learning_rate": 0.000188036410923277, "loss": 1.1284, "step": 126 }, { "epoch": 0.15, "grad_norm": 0.3887438178062439, "learning_rate": 0.00018777633289986998, "loss": 1.0324, "step": 127 }, { "epoch": 0.15, "grad_norm": 0.5829350352287292, "learning_rate": 0.00018751625487646294, "loss": 1.038, "step": 128 }, { "epoch": 0.15, "grad_norm": 0.5861855149269104, "learning_rate": 0.00018725617685305592, "loss": 0.9453, "step": 129 }, { "epoch": 0.15, "grad_norm": 0.40044718980789185, "learning_rate": 0.0001869960988296489, "loss": 1.1971, "step": 130 }, { "epoch": 0.15, "grad_norm": 0.4322098195552826, "learning_rate": 0.00018673602080624188, "loss": 0.9792, "step": 131 }, { "epoch": 0.16, "grad_norm": 0.4132952392101288, "learning_rate": 0.00018647594278283487, "loss": 0.9597, "step": 132 }, { "epoch": 0.16, "grad_norm": 0.4141194224357605, "learning_rate": 0.00018621586475942782, "loss": 1.0785, "step": 133 }, { "epoch": 0.16, "grad_norm": 0.5526332855224609, "learning_rate": 0.0001859557867360208, "loss": 1.0762, "step": 134 }, { "epoch": 0.16, "grad_norm": 0.6953147053718567, "learning_rate": 0.00018569570871261379, "loss": 1.1048, "step": 135 }, { "epoch": 0.16, "grad_norm": 0.4306538701057434, "learning_rate": 0.00018543563068920677, "loss": 0.9798, "step": 136 }, { "epoch": 0.16, "grad_norm": 0.32605746388435364, "learning_rate": 0.00018517555266579975, "loss": 0.9825, "step": 137 }, { "epoch": 0.16, "grad_norm": 0.48711439967155457, "learning_rate": 0.0001849154746423927, "loss": 0.9685, "step": 138 }, { "epoch": 0.16, "grad_norm": 0.7728849649429321, "learning_rate": 0.0001846553966189857, "loss": 1.2843, "step": 139 }, { "epoch": 0.16, "grad_norm": 0.38121744990348816, "learning_rate": 0.00018439531859557867, "loss": 0.8289, "step": 140 }, { "epoch": 0.17, "grad_norm": 0.5053502321243286, "learning_rate": 0.00018413524057217165, "loss": 1.0824, "step": 141 }, { "epoch": 0.17, "grad_norm": 0.44003796577453613, "learning_rate": 0.00018387516254876463, "loss": 1.0957, "step": 142 }, { "epoch": 0.17, "grad_norm": 0.35713884234428406, "learning_rate": 0.00018361508452535762, "loss": 0.7869, "step": 143 }, { "epoch": 0.17, "grad_norm": 0.6612699627876282, "learning_rate": 0.0001833550065019506, "loss": 1.1621, "step": 144 }, { "epoch": 0.17, "grad_norm": 0.4270091950893402, "learning_rate": 0.00018309492847854358, "loss": 0.8964, "step": 145 }, { "epoch": 0.17, "grad_norm": 0.42207202315330505, "learning_rate": 0.00018283485045513656, "loss": 1.1403, "step": 146 }, { "epoch": 0.17, "grad_norm": 0.48486125469207764, "learning_rate": 0.00018257477243172954, "loss": 0.9993, "step": 147 }, { "epoch": 0.17, "grad_norm": 0.4205480217933655, "learning_rate": 0.0001823146944083225, "loss": 1.0404, "step": 148 }, { "epoch": 0.18, "grad_norm": 0.3420919179916382, "learning_rate": 0.00018205461638491548, "loss": 1.1434, "step": 149 }, { "epoch": 0.18, "grad_norm": 0.6411290168762207, "learning_rate": 0.00018179453836150846, "loss": 1.2055, "step": 150 }, { "epoch": 0.18, "grad_norm": 0.7193886637687683, "learning_rate": 0.00018153446033810145, "loss": 0.9436, "step": 151 }, { "epoch": 0.18, "grad_norm": 0.34661969542503357, "learning_rate": 0.00018127438231469443, "loss": 0.8158, "step": 152 }, { "epoch": 0.18, "grad_norm": 0.4028843939304352, "learning_rate": 0.00018101430429128738, "loss": 0.9863, "step": 153 }, { "epoch": 0.18, "grad_norm": 0.3898251950740814, "learning_rate": 0.00018075422626788037, "loss": 0.9215, "step": 154 }, { "epoch": 0.18, "grad_norm": 0.3827955424785614, "learning_rate": 0.00018049414824447335, "loss": 0.9513, "step": 155 }, { "epoch": 0.18, "grad_norm": 0.3840239644050598, "learning_rate": 0.00018023407022106633, "loss": 0.8303, "step": 156 }, { "epoch": 0.18, "grad_norm": 0.42413073778152466, "learning_rate": 0.0001799739921976593, "loss": 0.8909, "step": 157 }, { "epoch": 0.19, "grad_norm": 0.47216784954071045, "learning_rate": 0.00017971391417425227, "loss": 1.0413, "step": 158 }, { "epoch": 0.19, "grad_norm": 0.35447853803634644, "learning_rate": 0.00017945383615084525, "loss": 0.9781, "step": 159 }, { "epoch": 0.19, "grad_norm": 0.46554017066955566, "learning_rate": 0.00017919375812743823, "loss": 0.7186, "step": 160 }, { "epoch": 0.19, "grad_norm": 0.3696407973766327, "learning_rate": 0.00017893368010403121, "loss": 1.0137, "step": 161 }, { "epoch": 0.19, "grad_norm": 0.5633952617645264, "learning_rate": 0.0001786736020806242, "loss": 1.0944, "step": 162 }, { "epoch": 0.19, "grad_norm": 0.3334963917732239, "learning_rate": 0.00017841352405721715, "loss": 1.2606, "step": 163 }, { "epoch": 0.19, "grad_norm": 0.4501505494117737, "learning_rate": 0.00017815344603381013, "loss": 0.7306, "step": 164 }, { "epoch": 0.19, "grad_norm": 0.44553884863853455, "learning_rate": 0.00017789336801040312, "loss": 0.871, "step": 165 }, { "epoch": 0.2, "grad_norm": 0.474342405796051, "learning_rate": 0.0001776332899869961, "loss": 0.7593, "step": 166 }, { "epoch": 0.2, "grad_norm": 0.7258349061012268, "learning_rate": 0.00017737321196358908, "loss": 1.4035, "step": 167 }, { "epoch": 0.2, "grad_norm": 0.43586501479148865, "learning_rate": 0.00017711313394018206, "loss": 1.1679, "step": 168 }, { "epoch": 0.2, "grad_norm": 0.4580634832382202, "learning_rate": 0.00017685305591677504, "loss": 0.9496, "step": 169 }, { "epoch": 0.2, "grad_norm": 0.40509316325187683, "learning_rate": 0.00017659297789336803, "loss": 0.9515, "step": 170 }, { "epoch": 0.2, "grad_norm": 0.3496972620487213, "learning_rate": 0.000176332899869961, "loss": 1.1846, "step": 171 }, { "epoch": 0.2, "grad_norm": 0.6468595862388611, "learning_rate": 0.000176072821846554, "loss": 1.0307, "step": 172 }, { "epoch": 0.2, "grad_norm": 0.3840126693248749, "learning_rate": 0.00017581274382314695, "loss": 1.0248, "step": 173 }, { "epoch": 0.2, "grad_norm": 0.4715554118156433, "learning_rate": 0.00017555266579973993, "loss": 1.1017, "step": 174 }, { "epoch": 0.21, "grad_norm": 0.38534194231033325, "learning_rate": 0.0001752925877763329, "loss": 1.0073, "step": 175 }, { "epoch": 0.21, "grad_norm": 0.3881916403770447, "learning_rate": 0.0001750325097529259, "loss": 0.8961, "step": 176 }, { "epoch": 0.21, "grad_norm": 0.49220091104507446, "learning_rate": 0.00017477243172951887, "loss": 1.0184, "step": 177 }, { "epoch": 0.21, "grad_norm": 0.38170936703681946, "learning_rate": 0.00017451235370611186, "loss": 1.0939, "step": 178 }, { "epoch": 0.21, "grad_norm": 0.3692156672477722, "learning_rate": 0.0001742522756827048, "loss": 1.085, "step": 179 }, { "epoch": 0.21, "grad_norm": 0.43161171674728394, "learning_rate": 0.0001739921976592978, "loss": 0.9098, "step": 180 }, { "epoch": 0.21, "grad_norm": 0.3874380886554718, "learning_rate": 0.00017373211963589078, "loss": 1.0108, "step": 181 }, { "epoch": 0.21, "grad_norm": 0.37556958198547363, "learning_rate": 0.00017347204161248376, "loss": 1.0344, "step": 182 }, { "epoch": 0.22, "grad_norm": 0.4288795292377472, "learning_rate": 0.00017321196358907674, "loss": 0.9329, "step": 183 }, { "epoch": 0.22, "grad_norm": 0.6587222218513489, "learning_rate": 0.0001729518855656697, "loss": 1.0312, "step": 184 }, { "epoch": 0.22, "grad_norm": 0.6700361371040344, "learning_rate": 0.00017269180754226268, "loss": 1.035, "step": 185 }, { "epoch": 0.22, "grad_norm": 0.5208828449249268, "learning_rate": 0.00017243172951885566, "loss": 1.0334, "step": 186 }, { "epoch": 0.22, "grad_norm": 0.5774813294410706, "learning_rate": 0.00017217165149544864, "loss": 1.0366, "step": 187 }, { "epoch": 0.22, "grad_norm": 0.3492725193500519, "learning_rate": 0.00017191157347204162, "loss": 1.0225, "step": 188 }, { "epoch": 0.22, "grad_norm": 0.3714265525341034, "learning_rate": 0.00017165149544863458, "loss": 1.0355, "step": 189 }, { "epoch": 0.22, "grad_norm": 0.4425973892211914, "learning_rate": 0.00017139141742522756, "loss": 0.974, "step": 190 }, { "epoch": 0.22, "grad_norm": 0.43257299065589905, "learning_rate": 0.00017113133940182054, "loss": 0.8614, "step": 191 }, { "epoch": 0.23, "grad_norm": 0.37084296345710754, "learning_rate": 0.00017087126137841353, "loss": 0.7135, "step": 192 }, { "epoch": 0.23, "grad_norm": 0.6695109009742737, "learning_rate": 0.0001706111833550065, "loss": 0.7722, "step": 193 }, { "epoch": 0.23, "grad_norm": 0.5144183039665222, "learning_rate": 0.0001703511053315995, "loss": 0.9364, "step": 194 }, { "epoch": 0.23, "grad_norm": 0.58673095703125, "learning_rate": 0.00017009102730819247, "loss": 1.1141, "step": 195 }, { "epoch": 0.23, "grad_norm": 0.4033479392528534, "learning_rate": 0.00016983094928478545, "loss": 1.0002, "step": 196 }, { "epoch": 0.23, "grad_norm": 0.38111910223960876, "learning_rate": 0.00016957087126137844, "loss": 0.9318, "step": 197 }, { "epoch": 0.23, "grad_norm": 0.4174363613128662, "learning_rate": 0.00016931079323797142, "loss": 0.9381, "step": 198 }, { "epoch": 0.23, "grad_norm": 0.3726828098297119, "learning_rate": 0.00016905071521456437, "loss": 1.0405, "step": 199 }, { "epoch": 0.24, "grad_norm": 0.5414443612098694, "learning_rate": 0.00016879063719115736, "loss": 1.0667, "step": 200 }, { "epoch": 0.24, "eval_loss": 1.0012115240097046, "eval_runtime": 546.9226, "eval_samples_per_second": 2.761, "eval_steps_per_second": 1.38, "step": 200 }, { "epoch": 0.24, "grad_norm": 0.523889422416687, "learning_rate": 0.00016853055916775034, "loss": 1.0408, "step": 201 }, { "epoch": 0.24, "grad_norm": 0.47460106015205383, "learning_rate": 0.00016827048114434332, "loss": 1.0327, "step": 202 }, { "epoch": 0.24, "grad_norm": 0.5161871314048767, "learning_rate": 0.0001680104031209363, "loss": 0.9104, "step": 203 }, { "epoch": 0.24, "grad_norm": 0.3816511631011963, "learning_rate": 0.00016775032509752926, "loss": 0.6875, "step": 204 }, { "epoch": 0.24, "grad_norm": 0.3574850261211395, "learning_rate": 0.00016749024707412224, "loss": 0.8079, "step": 205 }, { "epoch": 0.24, "grad_norm": 0.37416210770606995, "learning_rate": 0.00016723016905071522, "loss": 0.7479, "step": 206 }, { "epoch": 0.24, "grad_norm": 0.648332417011261, "learning_rate": 0.0001669700910273082, "loss": 1.0868, "step": 207 }, { "epoch": 0.24, "grad_norm": 0.37010350823402405, "learning_rate": 0.0001667100130039012, "loss": 0.8634, "step": 208 }, { "epoch": 0.25, "grad_norm": 0.5206050872802734, "learning_rate": 0.00016644993498049414, "loss": 0.8827, "step": 209 }, { "epoch": 0.25, "grad_norm": 0.4001045525074005, "learning_rate": 0.00016618985695708712, "loss": 1.1029, "step": 210 }, { "epoch": 0.25, "grad_norm": 0.3948250114917755, "learning_rate": 0.0001659297789336801, "loss": 0.7248, "step": 211 }, { "epoch": 0.25, "grad_norm": 0.4833560287952423, "learning_rate": 0.0001656697009102731, "loss": 0.7379, "step": 212 }, { "epoch": 0.25, "grad_norm": 0.4951235055923462, "learning_rate": 0.00016540962288686607, "loss": 1.0003, "step": 213 }, { "epoch": 0.25, "grad_norm": 0.7607774138450623, "learning_rate": 0.00016514954486345903, "loss": 0.8599, "step": 214 }, { "epoch": 0.25, "grad_norm": 0.3604733347892761, "learning_rate": 0.000164889466840052, "loss": 1.0301, "step": 215 }, { "epoch": 0.25, "grad_norm": 0.3514990508556366, "learning_rate": 0.000164629388816645, "loss": 1.072, "step": 216 }, { "epoch": 0.26, "grad_norm": 0.8136312365531921, "learning_rate": 0.00016436931079323797, "loss": 1.0795, "step": 217 }, { "epoch": 0.26, "grad_norm": 0.4305706024169922, "learning_rate": 0.00016410923276983095, "loss": 0.8288, "step": 218 }, { "epoch": 0.26, "grad_norm": 0.4228506088256836, "learning_rate": 0.00016384915474642394, "loss": 0.863, "step": 219 }, { "epoch": 0.26, "grad_norm": 0.3303283751010895, "learning_rate": 0.00016358907672301692, "loss": 0.8992, "step": 220 }, { "epoch": 0.26, "grad_norm": 0.4840926229953766, "learning_rate": 0.0001633289986996099, "loss": 0.9309, "step": 221 }, { "epoch": 0.26, "grad_norm": 0.3939206600189209, "learning_rate": 0.00016306892067620288, "loss": 1.0456, "step": 222 }, { "epoch": 0.26, "grad_norm": 0.4611961245536804, "learning_rate": 0.00016280884265279587, "loss": 1.1839, "step": 223 }, { "epoch": 0.26, "grad_norm": 0.4443643093109131, "learning_rate": 0.00016254876462938882, "loss": 0.8566, "step": 224 }, { "epoch": 0.27, "grad_norm": 0.38910534977912903, "learning_rate": 0.0001622886866059818, "loss": 0.9426, "step": 225 }, { "epoch": 0.27, "grad_norm": 0.6585952043533325, "learning_rate": 0.00016202860858257478, "loss": 1.0632, "step": 226 }, { "epoch": 0.27, "grad_norm": 0.2883739173412323, "learning_rate": 0.00016176853055916777, "loss": 0.8594, "step": 227 }, { "epoch": 0.27, "grad_norm": 0.4793940484523773, "learning_rate": 0.00016150845253576075, "loss": 0.9258, "step": 228 }, { "epoch": 0.27, "grad_norm": 0.45249128341674805, "learning_rate": 0.0001612483745123537, "loss": 0.875, "step": 229 }, { "epoch": 0.27, "grad_norm": 0.7601549029350281, "learning_rate": 0.00016098829648894669, "loss": 1.0903, "step": 230 }, { "epoch": 0.27, "grad_norm": 0.4474358856678009, "learning_rate": 0.00016072821846553967, "loss": 0.8636, "step": 231 }, { "epoch": 0.27, "grad_norm": 0.4709658920764923, "learning_rate": 0.00016046814044213265, "loss": 1.01, "step": 232 }, { "epoch": 0.27, "grad_norm": 0.49074116349220276, "learning_rate": 0.00016020806241872563, "loss": 1.129, "step": 233 }, { "epoch": 0.28, "grad_norm": 0.40011027455329895, "learning_rate": 0.0001599479843953186, "loss": 0.9789, "step": 234 }, { "epoch": 0.28, "grad_norm": 0.36765021085739136, "learning_rate": 0.00015968790637191157, "loss": 0.9647, "step": 235 }, { "epoch": 0.28, "grad_norm": 0.4226991832256317, "learning_rate": 0.00015942782834850455, "loss": 1.0593, "step": 236 }, { "epoch": 0.28, "grad_norm": 0.5633476376533508, "learning_rate": 0.00015916775032509753, "loss": 1.139, "step": 237 }, { "epoch": 0.28, "grad_norm": 0.4227031171321869, "learning_rate": 0.00015890767230169052, "loss": 0.9766, "step": 238 }, { "epoch": 0.28, "grad_norm": 0.7903754711151123, "learning_rate": 0.00015864759427828347, "loss": 1.0215, "step": 239 }, { "epoch": 0.28, "grad_norm": 0.5319734811782837, "learning_rate": 0.00015838751625487645, "loss": 1.0351, "step": 240 }, { "epoch": 0.28, "grad_norm": 0.38870692253112793, "learning_rate": 0.00015812743823146944, "loss": 1.1089, "step": 241 }, { "epoch": 0.29, "grad_norm": 0.5061522126197815, "learning_rate": 0.00015786736020806242, "loss": 0.9956, "step": 242 }, { "epoch": 0.29, "grad_norm": 0.6587597131729126, "learning_rate": 0.0001576072821846554, "loss": 0.9433, "step": 243 }, { "epoch": 0.29, "grad_norm": 0.500415027141571, "learning_rate": 0.00015734720416124838, "loss": 0.9071, "step": 244 }, { "epoch": 0.29, "grad_norm": 0.49865788221359253, "learning_rate": 0.00015708712613784136, "loss": 1.013, "step": 245 }, { "epoch": 0.29, "grad_norm": 0.472517192363739, "learning_rate": 0.00015682704811443435, "loss": 1.0552, "step": 246 }, { "epoch": 0.29, "grad_norm": 0.5091108083724976, "learning_rate": 0.00015656697009102733, "loss": 0.9919, "step": 247 }, { "epoch": 0.29, "grad_norm": 0.39484265446662903, "learning_rate": 0.0001563068920676203, "loss": 0.9945, "step": 248 }, { "epoch": 0.29, "grad_norm": 0.47099944949150085, "learning_rate": 0.00015604681404421327, "loss": 0.847, "step": 249 }, { "epoch": 0.29, "grad_norm": 0.4681323766708374, "learning_rate": 0.00015578673602080625, "loss": 0.8998, "step": 250 }, { "epoch": 0.3, "grad_norm": 0.35557135939598083, "learning_rate": 0.00015552665799739923, "loss": 0.7108, "step": 251 }, { "epoch": 0.3, "grad_norm": 0.7873208522796631, "learning_rate": 0.0001552665799739922, "loss": 1.1322, "step": 252 }, { "epoch": 0.3, "grad_norm": 0.45806723833084106, "learning_rate": 0.0001550065019505852, "loss": 0.9095, "step": 253 }, { "epoch": 0.3, "grad_norm": 0.639227032661438, "learning_rate": 0.00015474642392717815, "loss": 1.1761, "step": 254 }, { "epoch": 0.3, "grad_norm": 0.6849457621574402, "learning_rate": 0.00015448634590377113, "loss": 1.1391, "step": 255 }, { "epoch": 0.3, "grad_norm": 0.5833932757377625, "learning_rate": 0.00015422626788036411, "loss": 1.0434, "step": 256 }, { "epoch": 0.3, "grad_norm": 0.684618353843689, "learning_rate": 0.0001539661898569571, "loss": 0.9719, "step": 257 }, { "epoch": 0.3, "grad_norm": 0.4258657991886139, "learning_rate": 0.00015370611183355008, "loss": 1.1534, "step": 258 }, { "epoch": 0.31, "grad_norm": 0.5763025879859924, "learning_rate": 0.00015344603381014303, "loss": 0.9974, "step": 259 }, { "epoch": 0.31, "grad_norm": 0.5415167808532715, "learning_rate": 0.00015318595578673602, "loss": 1.0696, "step": 260 }, { "epoch": 0.31, "grad_norm": 0.43196481466293335, "learning_rate": 0.000152925877763329, "loss": 0.8191, "step": 261 }, { "epoch": 0.31, "grad_norm": 0.3914147913455963, "learning_rate": 0.00015266579973992198, "loss": 0.8465, "step": 262 }, { "epoch": 0.31, "grad_norm": 0.4838383197784424, "learning_rate": 0.00015240572171651496, "loss": 1.0027, "step": 263 }, { "epoch": 0.31, "grad_norm": 0.3581622540950775, "learning_rate": 0.00015214564369310792, "loss": 1.1247, "step": 264 }, { "epoch": 0.31, "grad_norm": 0.9383752942085266, "learning_rate": 0.0001518855656697009, "loss": 1.2207, "step": 265 }, { "epoch": 0.31, "grad_norm": 0.5058914422988892, "learning_rate": 0.00015162548764629388, "loss": 1.0036, "step": 266 }, { "epoch": 0.31, "grad_norm": 0.4513028860092163, "learning_rate": 0.0001513654096228869, "loss": 1.0949, "step": 267 }, { "epoch": 0.32, "grad_norm": 0.5779305696487427, "learning_rate": 0.00015110533159947987, "loss": 1.0035, "step": 268 }, { "epoch": 0.32, "grad_norm": 0.6758905649185181, "learning_rate": 0.00015084525357607283, "loss": 1.0163, "step": 269 }, { "epoch": 0.32, "grad_norm": 0.5109132528305054, "learning_rate": 0.0001505851755526658, "loss": 0.7357, "step": 270 }, { "epoch": 0.32, "grad_norm": 0.42248305678367615, "learning_rate": 0.0001503250975292588, "loss": 1.0664, "step": 271 }, { "epoch": 0.32, "grad_norm": 0.7441749572753906, "learning_rate": 0.00015006501950585178, "loss": 0.9483, "step": 272 }, { "epoch": 0.32, "grad_norm": 0.5364653468132019, "learning_rate": 0.00014980494148244476, "loss": 1.1671, "step": 273 }, { "epoch": 0.32, "grad_norm": 0.3390597701072693, "learning_rate": 0.0001495448634590377, "loss": 0.9616, "step": 274 }, { "epoch": 0.32, "grad_norm": 0.3810036778450012, "learning_rate": 0.0001492847854356307, "loss": 0.9419, "step": 275 }, { "epoch": 0.33, "grad_norm": 0.39790648221969604, "learning_rate": 0.00014902470741222368, "loss": 0.947, "step": 276 }, { "epoch": 0.33, "grad_norm": 0.5461649894714355, "learning_rate": 0.00014876462938881666, "loss": 1.0065, "step": 277 }, { "epoch": 0.33, "grad_norm": 0.376022607088089, "learning_rate": 0.00014850455136540964, "loss": 0.8626, "step": 278 }, { "epoch": 0.33, "grad_norm": 0.47341254353523254, "learning_rate": 0.0001482444733420026, "loss": 0.9812, "step": 279 }, { "epoch": 0.33, "grad_norm": 0.5151678323745728, "learning_rate": 0.00014798439531859558, "loss": 0.9688, "step": 280 }, { "epoch": 0.33, "grad_norm": 0.4801676273345947, "learning_rate": 0.00014772431729518856, "loss": 0.9348, "step": 281 }, { "epoch": 0.33, "grad_norm": 0.43506500124931335, "learning_rate": 0.00014746423927178154, "loss": 1.0792, "step": 282 }, { "epoch": 0.33, "grad_norm": 0.5347827076911926, "learning_rate": 0.00014720416124837452, "loss": 1.1285, "step": 283 }, { "epoch": 0.33, "grad_norm": 0.5323308706283569, "learning_rate": 0.00014694408322496748, "loss": 1.047, "step": 284 }, { "epoch": 0.34, "grad_norm": 0.364046573638916, "learning_rate": 0.00014668400520156046, "loss": 1.0206, "step": 285 }, { "epoch": 0.34, "grad_norm": 0.41780295968055725, "learning_rate": 0.00014642392717815344, "loss": 1.011, "step": 286 }, { "epoch": 0.34, "grad_norm": 0.6010937690734863, "learning_rate": 0.00014616384915474643, "loss": 0.9836, "step": 287 }, { "epoch": 0.34, "grad_norm": 0.3574303090572357, "learning_rate": 0.0001459037711313394, "loss": 0.9051, "step": 288 }, { "epoch": 0.34, "grad_norm": 0.3346600830554962, "learning_rate": 0.0001456436931079324, "loss": 1.0333, "step": 289 }, { "epoch": 0.34, "grad_norm": 0.48559319972991943, "learning_rate": 0.00014538361508452537, "loss": 0.8363, "step": 290 }, { "epoch": 0.34, "grad_norm": 0.4872112274169922, "learning_rate": 0.00014512353706111836, "loss": 0.9565, "step": 291 }, { "epoch": 0.34, "grad_norm": 0.37863266468048096, "learning_rate": 0.00014486345903771134, "loss": 0.9869, "step": 292 }, { "epoch": 0.35, "grad_norm": 0.6248475909233093, "learning_rate": 0.00014460338101430432, "loss": 1.1117, "step": 293 }, { "epoch": 0.35, "grad_norm": 0.40380629897117615, "learning_rate": 0.00014434330299089727, "loss": 0.895, "step": 294 }, { "epoch": 0.35, "grad_norm": 0.4236217439174652, "learning_rate": 0.00014408322496749026, "loss": 0.8849, "step": 295 }, { "epoch": 0.35, "grad_norm": 0.35017380118370056, "learning_rate": 0.00014382314694408324, "loss": 0.8472, "step": 296 }, { "epoch": 0.35, "grad_norm": 0.5927186608314514, "learning_rate": 0.00014356306892067622, "loss": 1.0073, "step": 297 }, { "epoch": 0.35, "grad_norm": 0.7567859292030334, "learning_rate": 0.0001433029908972692, "loss": 0.9784, "step": 298 }, { "epoch": 0.35, "grad_norm": 0.4356285035610199, "learning_rate": 0.00014304291287386216, "loss": 0.8339, "step": 299 }, { "epoch": 0.35, "grad_norm": 0.3144444525241852, "learning_rate": 0.00014278283485045514, "loss": 1.0751, "step": 300 }, { "epoch": 0.35, "eval_loss": 0.9849200248718262, "eval_runtime": 546.1357, "eval_samples_per_second": 2.765, "eval_steps_per_second": 1.382, "step": 300 }, { "epoch": 0.35, "grad_norm": 0.5712672472000122, "learning_rate": 0.00014252275682704812, "loss": 0.9618, "step": 301 }, { "epoch": 0.36, "grad_norm": 0.7324238419532776, "learning_rate": 0.0001422626788036411, "loss": 1.1554, "step": 302 }, { "epoch": 0.36, "grad_norm": 0.4014214277267456, "learning_rate": 0.0001420026007802341, "loss": 0.9471, "step": 303 }, { "epoch": 0.36, "grad_norm": 0.3690221905708313, "learning_rate": 0.00014174252275682704, "loss": 0.691, "step": 304 }, { "epoch": 0.36, "grad_norm": 0.3683489263057709, "learning_rate": 0.00014148244473342002, "loss": 0.7806, "step": 305 }, { "epoch": 0.36, "grad_norm": 0.6884939670562744, "learning_rate": 0.000141222366710013, "loss": 1.1163, "step": 306 }, { "epoch": 0.36, "grad_norm": 0.5952924489974976, "learning_rate": 0.000140962288686606, "loss": 1.0318, "step": 307 }, { "epoch": 0.36, "grad_norm": 0.4473588764667511, "learning_rate": 0.00014070221066319897, "loss": 1.0514, "step": 308 }, { "epoch": 0.36, "grad_norm": 0.3944532871246338, "learning_rate": 0.00014044213263979193, "loss": 0.8776, "step": 309 }, { "epoch": 0.37, "grad_norm": 0.6479160189628601, "learning_rate": 0.0001401820546163849, "loss": 1.0586, "step": 310 }, { "epoch": 0.37, "grad_norm": 0.4719555974006653, "learning_rate": 0.0001399219765929779, "loss": 1.1372, "step": 311 }, { "epoch": 0.37, "grad_norm": 0.4791443347930908, "learning_rate": 0.00013966189856957087, "loss": 1.0131, "step": 312 }, { "epoch": 0.37, "grad_norm": 0.6476565599441528, "learning_rate": 0.00013940182054616385, "loss": 0.9054, "step": 313 }, { "epoch": 0.37, "grad_norm": 0.4465392529964447, "learning_rate": 0.00013914174252275684, "loss": 1.09, "step": 314 }, { "epoch": 0.37, "grad_norm": 0.4066149890422821, "learning_rate": 0.00013888166449934982, "loss": 0.8515, "step": 315 }, { "epoch": 0.37, "grad_norm": 0.3584599196910858, "learning_rate": 0.0001386215864759428, "loss": 0.8008, "step": 316 }, { "epoch": 0.37, "grad_norm": 0.6454095840454102, "learning_rate": 0.00013836150845253578, "loss": 1.1542, "step": 317 }, { "epoch": 0.37, "grad_norm": 0.6096594333648682, "learning_rate": 0.00013810143042912877, "loss": 1.26, "step": 318 }, { "epoch": 0.38, "grad_norm": 0.42694196105003357, "learning_rate": 0.00013784135240572172, "loss": 0.754, "step": 319 }, { "epoch": 0.38, "grad_norm": 0.5280473232269287, "learning_rate": 0.0001375812743823147, "loss": 1.0522, "step": 320 }, { "epoch": 0.38, "grad_norm": 0.46211832761764526, "learning_rate": 0.00013732119635890769, "loss": 1.094, "step": 321 }, { "epoch": 0.38, "grad_norm": 0.3849136531352997, "learning_rate": 0.00013706111833550067, "loss": 0.9067, "step": 322 }, { "epoch": 0.38, "grad_norm": 0.37398576736450195, "learning_rate": 0.00013680104031209365, "loss": 0.9948, "step": 323 }, { "epoch": 0.38, "grad_norm": 0.38779109716415405, "learning_rate": 0.0001365409622886866, "loss": 1.0286, "step": 324 }, { "epoch": 0.38, "grad_norm": 0.4133855700492859, "learning_rate": 0.0001362808842652796, "loss": 0.977, "step": 325 }, { "epoch": 0.38, "grad_norm": 0.44233980774879456, "learning_rate": 0.00013602080624187257, "loss": 0.908, "step": 326 }, { "epoch": 0.39, "grad_norm": 0.4096635580062866, "learning_rate": 0.00013576072821846555, "loss": 0.88, "step": 327 }, { "epoch": 0.39, "grad_norm": 0.4307350814342499, "learning_rate": 0.00013550065019505853, "loss": 0.8848, "step": 328 }, { "epoch": 0.39, "grad_norm": 0.33225497603416443, "learning_rate": 0.0001352405721716515, "loss": 0.8369, "step": 329 }, { "epoch": 0.39, "grad_norm": 0.5546039342880249, "learning_rate": 0.00013498049414824447, "loss": 0.6756, "step": 330 }, { "epoch": 0.39, "grad_norm": 0.4454500377178192, "learning_rate": 0.00013472041612483745, "loss": 0.994, "step": 331 }, { "epoch": 0.39, "grad_norm": 0.37814733386039734, "learning_rate": 0.00013446033810143044, "loss": 1.028, "step": 332 }, { "epoch": 0.39, "grad_norm": 0.8602733612060547, "learning_rate": 0.00013420026007802342, "loss": 1.0618, "step": 333 }, { "epoch": 0.39, "grad_norm": 0.5205318331718445, "learning_rate": 0.00013394018205461637, "loss": 0.9995, "step": 334 }, { "epoch": 0.39, "grad_norm": 0.3141395151615143, "learning_rate": 0.00013368010403120935, "loss": 1.0196, "step": 335 }, { "epoch": 0.4, "grad_norm": 0.3242855668067932, "learning_rate": 0.00013342002600780234, "loss": 0.9072, "step": 336 }, { "epoch": 0.4, "grad_norm": 0.37035492062568665, "learning_rate": 0.00013315994798439532, "loss": 0.8574, "step": 337 }, { "epoch": 0.4, "grad_norm": 0.46976369619369507, "learning_rate": 0.0001328998699609883, "loss": 1.0797, "step": 338 }, { "epoch": 0.4, "grad_norm": 0.4217787981033325, "learning_rate": 0.00013263979193758128, "loss": 0.9538, "step": 339 }, { "epoch": 0.4, "grad_norm": 0.7571201920509338, "learning_rate": 0.00013237971391417427, "loss": 1.0524, "step": 340 }, { "epoch": 0.4, "grad_norm": 0.7066246271133423, "learning_rate": 0.00013211963589076725, "loss": 0.9775, "step": 341 }, { "epoch": 0.4, "grad_norm": 0.3496040403842926, "learning_rate": 0.00013185955786736023, "loss": 0.9926, "step": 342 }, { "epoch": 0.4, "grad_norm": 0.49747544527053833, "learning_rate": 0.0001315994798439532, "loss": 1.0803, "step": 343 }, { "epoch": 0.41, "grad_norm": 0.406820148229599, "learning_rate": 0.00013133940182054617, "loss": 0.8144, "step": 344 }, { "epoch": 0.41, "grad_norm": 0.3256113827228546, "learning_rate": 0.00013107932379713915, "loss": 0.8676, "step": 345 }, { "epoch": 0.41, "grad_norm": 0.48141512274742126, "learning_rate": 0.00013081924577373213, "loss": 0.9067, "step": 346 }, { "epoch": 0.41, "grad_norm": 0.35172903537750244, "learning_rate": 0.0001305591677503251, "loss": 0.9535, "step": 347 }, { "epoch": 0.41, "grad_norm": 0.39796268939971924, "learning_rate": 0.0001302990897269181, "loss": 1.0817, "step": 348 }, { "epoch": 0.41, "grad_norm": 0.4287821650505066, "learning_rate": 0.00013003901170351105, "loss": 0.9363, "step": 349 }, { "epoch": 0.41, "grad_norm": 0.45490118861198425, "learning_rate": 0.00012977893368010403, "loss": 1.0131, "step": 350 }, { "epoch": 0.41, "grad_norm": 0.4730069935321808, "learning_rate": 0.00012951885565669702, "loss": 1.0037, "step": 351 }, { "epoch": 0.41, "grad_norm": 0.34408411383628845, "learning_rate": 0.00012925877763329, "loss": 0.737, "step": 352 }, { "epoch": 0.42, "grad_norm": 0.5118675827980042, "learning_rate": 0.00012899869960988298, "loss": 1.0871, "step": 353 }, { "epoch": 0.42, "grad_norm": 0.45648622512817383, "learning_rate": 0.00012873862158647593, "loss": 0.8413, "step": 354 }, { "epoch": 0.42, "grad_norm": 0.30808356404304504, "learning_rate": 0.00012847854356306892, "loss": 0.7535, "step": 355 }, { "epoch": 0.42, "grad_norm": 0.4459214508533478, "learning_rate": 0.0001282184655396619, "loss": 0.9316, "step": 356 }, { "epoch": 0.42, "grad_norm": 0.7930001020431519, "learning_rate": 0.00012795838751625488, "loss": 0.9959, "step": 357 }, { "epoch": 0.42, "grad_norm": 0.5756102204322815, "learning_rate": 0.00012769830949284786, "loss": 1.0373, "step": 358 }, { "epoch": 0.42, "grad_norm": 0.37661170959472656, "learning_rate": 0.00012743823146944082, "loss": 0.885, "step": 359 }, { "epoch": 0.42, "grad_norm": 0.5622540712356567, "learning_rate": 0.0001271781534460338, "loss": 0.8901, "step": 360 }, { "epoch": 0.43, "grad_norm": 0.42885056138038635, "learning_rate": 0.00012691807542262678, "loss": 0.9577, "step": 361 }, { "epoch": 0.43, "grad_norm": 0.38707178831100464, "learning_rate": 0.00012665799739921976, "loss": 0.9606, "step": 362 }, { "epoch": 0.43, "grad_norm": 0.3369719684123993, "learning_rate": 0.00012639791937581275, "loss": 0.8205, "step": 363 }, { "epoch": 0.43, "grad_norm": 0.35598838329315186, "learning_rate": 0.00012613784135240573, "loss": 0.7119, "step": 364 }, { "epoch": 0.43, "grad_norm": 0.6167771816253662, "learning_rate": 0.0001258777633289987, "loss": 1.2462, "step": 365 }, { "epoch": 0.43, "grad_norm": 0.3976896107196808, "learning_rate": 0.0001256176853055917, "loss": 0.981, "step": 366 }, { "epoch": 0.43, "grad_norm": 0.4538115859031677, "learning_rate": 0.00012535760728218468, "loss": 1.0327, "step": 367 }, { "epoch": 0.43, "grad_norm": 0.7170857787132263, "learning_rate": 0.00012509752925877766, "loss": 0.9929, "step": 368 }, { "epoch": 0.43, "grad_norm": 0.45370182394981384, "learning_rate": 0.00012483745123537064, "loss": 0.7871, "step": 369 }, { "epoch": 0.44, "grad_norm": 0.4275226891040802, "learning_rate": 0.0001245773732119636, "loss": 0.7853, "step": 370 }, { "epoch": 0.44, "grad_norm": 0.4282675087451935, "learning_rate": 0.00012431729518855658, "loss": 1.0487, "step": 371 }, { "epoch": 0.44, "grad_norm": 0.6588276624679565, "learning_rate": 0.00012405721716514956, "loss": 0.9411, "step": 372 }, { "epoch": 0.44, "grad_norm": 0.6060877442359924, "learning_rate": 0.00012379713914174254, "loss": 0.9601, "step": 373 }, { "epoch": 0.44, "grad_norm": 0.5504083037376404, "learning_rate": 0.00012353706111833552, "loss": 0.9798, "step": 374 }, { "epoch": 0.44, "grad_norm": 0.40328875184059143, "learning_rate": 0.00012327698309492848, "loss": 0.9134, "step": 375 }, { "epoch": 0.44, "grad_norm": 0.5537497401237488, "learning_rate": 0.00012301690507152146, "loss": 0.8478, "step": 376 }, { "epoch": 0.44, "grad_norm": 0.4089404046535492, "learning_rate": 0.00012275682704811444, "loss": 0.9013, "step": 377 }, { "epoch": 0.45, "grad_norm": 0.47143158316612244, "learning_rate": 0.00012249674902470743, "loss": 0.9138, "step": 378 }, { "epoch": 0.45, "grad_norm": 0.7318865060806274, "learning_rate": 0.0001222366710013004, "loss": 1.1403, "step": 379 }, { "epoch": 0.45, "grad_norm": 0.537233293056488, "learning_rate": 0.00012197659297789336, "loss": 0.9555, "step": 380 }, { "epoch": 0.45, "grad_norm": 0.3424953520298004, "learning_rate": 0.00012171651495448635, "loss": 0.8932, "step": 381 }, { "epoch": 0.45, "grad_norm": 0.33310917019844055, "learning_rate": 0.00012145643693107933, "loss": 1.0403, "step": 382 }, { "epoch": 0.45, "grad_norm": 0.46227535605430603, "learning_rate": 0.00012119635890767231, "loss": 0.9926, "step": 383 }, { "epoch": 0.45, "grad_norm": 0.5558334589004517, "learning_rate": 0.00012093628088426529, "loss": 0.9385, "step": 384 }, { "epoch": 0.45, "grad_norm": 0.648505449295044, "learning_rate": 0.00012067620286085826, "loss": 1.2165, "step": 385 }, { "epoch": 0.45, "grad_norm": 0.4283152222633362, "learning_rate": 0.00012041612483745124, "loss": 0.8768, "step": 386 }, { "epoch": 0.46, "grad_norm": 0.38274627923965454, "learning_rate": 0.00012015604681404422, "loss": 0.8584, "step": 387 }, { "epoch": 0.46, "grad_norm": 0.3748539090156555, "learning_rate": 0.0001198959687906372, "loss": 0.951, "step": 388 }, { "epoch": 0.46, "grad_norm": 0.509125828742981, "learning_rate": 0.00011963589076723019, "loss": 0.9077, "step": 389 }, { "epoch": 0.46, "grad_norm": 0.4202275276184082, "learning_rate": 0.00011937581274382314, "loss": 0.9849, "step": 390 }, { "epoch": 0.46, "grad_norm": 0.4720480144023895, "learning_rate": 0.00011911573472041613, "loss": 1.2436, "step": 391 }, { "epoch": 0.46, "grad_norm": 0.365005761384964, "learning_rate": 0.00011885565669700911, "loss": 0.7927, "step": 392 }, { "epoch": 0.46, "grad_norm": 0.7129966616630554, "learning_rate": 0.00011859557867360209, "loss": 1.2607, "step": 393 }, { "epoch": 0.46, "grad_norm": 0.3527359068393707, "learning_rate": 0.00011833550065019507, "loss": 0.9059, "step": 394 }, { "epoch": 0.47, "grad_norm": 0.42210349440574646, "learning_rate": 0.00011807542262678804, "loss": 0.8577, "step": 395 }, { "epoch": 0.47, "grad_norm": 0.6976379752159119, "learning_rate": 0.00011781534460338102, "loss": 1.1473, "step": 396 }, { "epoch": 0.47, "grad_norm": 0.38397690653800964, "learning_rate": 0.000117555266579974, "loss": 0.8791, "step": 397 }, { "epoch": 0.47, "grad_norm": 0.42243918776512146, "learning_rate": 0.00011729518855656699, "loss": 1.0964, "step": 398 }, { "epoch": 0.47, "grad_norm": 0.41610145568847656, "learning_rate": 0.00011703511053315997, "loss": 0.7124, "step": 399 }, { "epoch": 0.47, "grad_norm": 0.4049556255340576, "learning_rate": 0.00011677503250975293, "loss": 0.8838, "step": 400 }, { "epoch": 0.47, "eval_loss": 0.9695621132850647, "eval_runtime": 547.2222, "eval_samples_per_second": 2.759, "eval_steps_per_second": 1.38, "step": 400 }, { "epoch": 0.47, "grad_norm": 0.5041651129722595, "learning_rate": 0.00011651495448634591, "loss": 1.1952, "step": 401 }, { "epoch": 0.47, "grad_norm": 0.4359878599643707, "learning_rate": 0.00011625487646293889, "loss": 1.0123, "step": 402 }, { "epoch": 0.47, "grad_norm": 0.5219016671180725, "learning_rate": 0.00011599479843953187, "loss": 0.9308, "step": 403 }, { "epoch": 0.48, "grad_norm": 0.5299083590507507, "learning_rate": 0.00011573472041612485, "loss": 0.8205, "step": 404 }, { "epoch": 0.48, "grad_norm": 0.3952549695968628, "learning_rate": 0.00011547464239271781, "loss": 0.7662, "step": 405 }, { "epoch": 0.48, "grad_norm": 0.4714783728122711, "learning_rate": 0.00011521456436931079, "loss": 0.8575, "step": 406 }, { "epoch": 0.48, "grad_norm": 0.33364737033843994, "learning_rate": 0.00011495448634590377, "loss": 1.0812, "step": 407 }, { "epoch": 0.48, "grad_norm": 0.3483280837535858, "learning_rate": 0.00011469440832249676, "loss": 0.8616, "step": 408 }, { "epoch": 0.48, "grad_norm": 0.45787307620048523, "learning_rate": 0.00011443433029908974, "loss": 0.9106, "step": 409 }, { "epoch": 0.48, "grad_norm": 0.37171471118927, "learning_rate": 0.0001141742522756827, "loss": 0.8991, "step": 410 }, { "epoch": 0.48, "grad_norm": 0.4171783924102783, "learning_rate": 0.00011391417425227569, "loss": 1.018, "step": 411 }, { "epoch": 0.49, "grad_norm": 0.33541813492774963, "learning_rate": 0.00011365409622886867, "loss": 0.9027, "step": 412 }, { "epoch": 0.49, "grad_norm": 0.3962589204311371, "learning_rate": 0.00011339401820546165, "loss": 0.9024, "step": 413 }, { "epoch": 0.49, "grad_norm": 0.4545789957046509, "learning_rate": 0.00011313394018205464, "loss": 0.9486, "step": 414 }, { "epoch": 0.49, "grad_norm": 0.3594336211681366, "learning_rate": 0.00011287386215864759, "loss": 1.01, "step": 415 }, { "epoch": 0.49, "grad_norm": 0.5460087060928345, "learning_rate": 0.00011261378413524057, "loss": 0.9236, "step": 416 }, { "epoch": 0.49, "grad_norm": 0.35831183195114136, "learning_rate": 0.00011235370611183355, "loss": 0.9973, "step": 417 }, { "epoch": 0.49, "grad_norm": 0.7852218747138977, "learning_rate": 0.00011209362808842654, "loss": 0.9471, "step": 418 }, { "epoch": 0.49, "grad_norm": 0.3702365458011627, "learning_rate": 0.00011183355006501952, "loss": 0.9709, "step": 419 }, { "epoch": 0.49, "grad_norm": 0.35481539368629456, "learning_rate": 0.00011157347204161249, "loss": 0.9658, "step": 420 }, { "epoch": 0.5, "grad_norm": 0.47527754306793213, "learning_rate": 0.00011131339401820547, "loss": 0.7476, "step": 421 }, { "epoch": 0.5, "grad_norm": 0.5047522187232971, "learning_rate": 0.00011105331599479845, "loss": 0.9477, "step": 422 }, { "epoch": 0.5, "grad_norm": 0.37565183639526367, "learning_rate": 0.00011079323797139143, "loss": 0.9696, "step": 423 }, { "epoch": 0.5, "grad_norm": 0.45552632212638855, "learning_rate": 0.00011053315994798442, "loss": 0.8612, "step": 424 }, { "epoch": 0.5, "grad_norm": 0.5363200306892395, "learning_rate": 0.00011027308192457737, "loss": 1.0803, "step": 425 }, { "epoch": 0.5, "grad_norm": 0.3728126883506775, "learning_rate": 0.00011001300390117035, "loss": 0.9001, "step": 426 }, { "epoch": 0.5, "grad_norm": 0.34863632917404175, "learning_rate": 0.00010975292587776334, "loss": 1.0043, "step": 427 }, { "epoch": 0.5, "grad_norm": 0.47629523277282715, "learning_rate": 0.00010949284785435632, "loss": 1.0766, "step": 428 }, { "epoch": 0.51, "grad_norm": 0.41906920075416565, "learning_rate": 0.0001092327698309493, "loss": 0.9736, "step": 429 }, { "epoch": 0.51, "grad_norm": 0.4399062395095825, "learning_rate": 0.00010897269180754226, "loss": 0.9656, "step": 430 }, { "epoch": 0.51, "grad_norm": 0.3949640691280365, "learning_rate": 0.00010871261378413524, "loss": 1.1373, "step": 431 }, { "epoch": 0.51, "grad_norm": 0.4008437395095825, "learning_rate": 0.00010845253576072822, "loss": 1.2063, "step": 432 }, { "epoch": 0.51, "grad_norm": 0.5856128334999084, "learning_rate": 0.0001081924577373212, "loss": 0.8815, "step": 433 }, { "epoch": 0.51, "grad_norm": 0.39964592456817627, "learning_rate": 0.00010793237971391418, "loss": 1.0336, "step": 434 }, { "epoch": 0.51, "grad_norm": 0.45366355776786804, "learning_rate": 0.00010767230169050715, "loss": 0.9597, "step": 435 }, { "epoch": 0.51, "grad_norm": 0.3872968256473541, "learning_rate": 0.00010741222366710013, "loss": 0.8672, "step": 436 }, { "epoch": 0.51, "grad_norm": 0.571480393409729, "learning_rate": 0.00010715214564369312, "loss": 0.8613, "step": 437 }, { "epoch": 0.52, "grad_norm": 0.3737768232822418, "learning_rate": 0.0001068920676202861, "loss": 0.6288, "step": 438 }, { "epoch": 0.52, "grad_norm": 0.4427306354045868, "learning_rate": 0.00010663198959687908, "loss": 0.682, "step": 439 }, { "epoch": 0.52, "grad_norm": 0.43474483489990234, "learning_rate": 0.00010637191157347204, "loss": 0.9893, "step": 440 }, { "epoch": 0.52, "grad_norm": 0.3701932728290558, "learning_rate": 0.00010611183355006502, "loss": 0.8146, "step": 441 }, { "epoch": 0.52, "grad_norm": 0.3685742914676666, "learning_rate": 0.000105851755526658, "loss": 0.8658, "step": 442 }, { "epoch": 0.52, "grad_norm": 0.4988309144973755, "learning_rate": 0.00010559167750325098, "loss": 0.8378, "step": 443 }, { "epoch": 0.52, "grad_norm": 0.4663272500038147, "learning_rate": 0.00010533159947984396, "loss": 0.8429, "step": 444 }, { "epoch": 0.52, "grad_norm": 0.4629034698009491, "learning_rate": 0.00010507152145643693, "loss": 1.0345, "step": 445 }, { "epoch": 0.53, "grad_norm": 0.4492078125476837, "learning_rate": 0.00010481144343302992, "loss": 0.9371, "step": 446 }, { "epoch": 0.53, "grad_norm": 0.6658744812011719, "learning_rate": 0.0001045513654096229, "loss": 1.0248, "step": 447 }, { "epoch": 0.53, "grad_norm": 0.4415484070777893, "learning_rate": 0.00010429128738621588, "loss": 0.8766, "step": 448 }, { "epoch": 0.53, "grad_norm": 0.6641436219215393, "learning_rate": 0.00010403120936280886, "loss": 1.0809, "step": 449 }, { "epoch": 0.53, "grad_norm": 0.43958601355552673, "learning_rate": 0.00010377113133940182, "loss": 0.6404, "step": 450 }, { "epoch": 0.53, "grad_norm": 0.3857196867465973, "learning_rate": 0.0001035110533159948, "loss": 0.992, "step": 451 }, { "epoch": 0.53, "grad_norm": 0.34215670824050903, "learning_rate": 0.00010325097529258778, "loss": 0.6554, "step": 452 }, { "epoch": 0.53, "grad_norm": 0.4773356318473816, "learning_rate": 0.00010299089726918076, "loss": 0.9528, "step": 453 }, { "epoch": 0.53, "grad_norm": 0.571839451789856, "learning_rate": 0.00010273081924577375, "loss": 0.8032, "step": 454 }, { "epoch": 0.54, "grad_norm": 0.42349740862846375, "learning_rate": 0.0001024707412223667, "loss": 0.9149, "step": 455 }, { "epoch": 0.54, "grad_norm": 0.398220032453537, "learning_rate": 0.00010221066319895968, "loss": 0.8322, "step": 456 }, { "epoch": 0.54, "grad_norm": 0.6505610942840576, "learning_rate": 0.00010195058517555267, "loss": 0.9302, "step": 457 }, { "epoch": 0.54, "grad_norm": 0.6050857305526733, "learning_rate": 0.00010169050715214565, "loss": 1.0134, "step": 458 }, { "epoch": 0.54, "grad_norm": 0.7780768871307373, "learning_rate": 0.00010143042912873863, "loss": 0.9798, "step": 459 }, { "epoch": 0.54, "grad_norm": 0.4089425802230835, "learning_rate": 0.0001011703511053316, "loss": 0.9293, "step": 460 }, { "epoch": 0.54, "grad_norm": 0.3921873867511749, "learning_rate": 0.00010091027308192458, "loss": 0.9423, "step": 461 }, { "epoch": 0.54, "grad_norm": 0.5735834240913391, "learning_rate": 0.00010065019505851756, "loss": 1.0119, "step": 462 }, { "epoch": 0.55, "grad_norm": 0.3620048463344574, "learning_rate": 0.00010039011703511055, "loss": 0.7793, "step": 463 }, { "epoch": 0.55, "grad_norm": 0.6324208974838257, "learning_rate": 0.00010013003901170353, "loss": 1.2616, "step": 464 }, { "epoch": 0.55, "grad_norm": 0.435024619102478, "learning_rate": 9.98699609882965e-05, "loss": 0.8547, "step": 465 }, { "epoch": 0.55, "grad_norm": 0.41158631443977356, "learning_rate": 9.960988296488946e-05, "loss": 0.9606, "step": 466 }, { "epoch": 0.55, "grad_norm": 0.40144023299217224, "learning_rate": 9.934980494148245e-05, "loss": 0.9881, "step": 467 }, { "epoch": 0.55, "grad_norm": 0.4548628628253937, "learning_rate": 9.908972691807543e-05, "loss": 1.0575, "step": 468 }, { "epoch": 0.55, "grad_norm": 0.47592610120773315, "learning_rate": 9.88296488946684e-05, "loss": 1.0998, "step": 469 }, { "epoch": 0.55, "grad_norm": 0.37015900015830994, "learning_rate": 9.856957087126138e-05, "loss": 0.8345, "step": 470 }, { "epoch": 0.55, "grad_norm": 0.44897329807281494, "learning_rate": 9.830949284785436e-05, "loss": 1.0702, "step": 471 }, { "epoch": 0.56, "grad_norm": 0.3390595614910126, "learning_rate": 9.804941482444734e-05, "loss": 0.8154, "step": 472 }, { "epoch": 0.56, "grad_norm": 0.36674201488494873, "learning_rate": 9.778933680104033e-05, "loss": 0.8116, "step": 473 }, { "epoch": 0.56, "grad_norm": 0.27573856711387634, "learning_rate": 9.75292587776333e-05, "loss": 0.8189, "step": 474 }, { "epoch": 0.56, "grad_norm": 0.4196186065673828, "learning_rate": 9.726918075422628e-05, "loss": 0.9585, "step": 475 }, { "epoch": 0.56, "grad_norm": 0.36988577246665955, "learning_rate": 9.700910273081925e-05, "loss": 0.8305, "step": 476 }, { "epoch": 0.56, "grad_norm": 0.49522173404693604, "learning_rate": 9.674902470741223e-05, "loss": 0.7897, "step": 477 }, { "epoch": 0.56, "grad_norm": 0.42384329438209534, "learning_rate": 9.648894668400521e-05, "loss": 0.8458, "step": 478 }, { "epoch": 0.56, "grad_norm": 0.6162276268005371, "learning_rate": 9.622886866059818e-05, "loss": 1.0375, "step": 479 }, { "epoch": 0.57, "grad_norm": 0.38178446888923645, "learning_rate": 9.596879063719116e-05, "loss": 0.9282, "step": 480 }, { "epoch": 0.57, "grad_norm": 0.4075917899608612, "learning_rate": 9.570871261378413e-05, "loss": 0.8738, "step": 481 }, { "epoch": 0.57, "grad_norm": 0.41484683752059937, "learning_rate": 9.544863459037711e-05, "loss": 0.9212, "step": 482 }, { "epoch": 0.57, "grad_norm": 0.43126848340034485, "learning_rate": 9.51885565669701e-05, "loss": 0.8831, "step": 483 }, { "epoch": 0.57, "grad_norm": 0.6840558648109436, "learning_rate": 9.492847854356308e-05, "loss": 0.8203, "step": 484 }, { "epoch": 0.57, "grad_norm": 0.5899393558502197, "learning_rate": 9.466840052015606e-05, "loss": 0.9899, "step": 485 }, { "epoch": 0.57, "grad_norm": 0.41661572456359863, "learning_rate": 9.440832249674903e-05, "loss": 1.1791, "step": 486 }, { "epoch": 0.57, "grad_norm": 0.473346084356308, "learning_rate": 9.414824447334201e-05, "loss": 0.9929, "step": 487 }, { "epoch": 0.57, "grad_norm": 0.361463725566864, "learning_rate": 9.388816644993499e-05, "loss": 0.9432, "step": 488 }, { "epoch": 0.58, "grad_norm": 0.3761877119541168, "learning_rate": 9.362808842652796e-05, "loss": 1.0647, "step": 489 }, { "epoch": 0.58, "grad_norm": 0.41097190976142883, "learning_rate": 9.336801040312094e-05, "loss": 0.8314, "step": 490 }, { "epoch": 0.58, "grad_norm": 0.3834340274333954, "learning_rate": 9.310793237971391e-05, "loss": 0.8198, "step": 491 }, { "epoch": 0.58, "grad_norm": 0.8354341387748718, "learning_rate": 9.284785435630689e-05, "loss": 0.7464, "step": 492 }, { "epoch": 0.58, "grad_norm": 0.657922089099884, "learning_rate": 9.258777633289987e-05, "loss": 1.0856, "step": 493 }, { "epoch": 0.58, "grad_norm": 0.450549453496933, "learning_rate": 9.232769830949284e-05, "loss": 1.1196, "step": 494 }, { "epoch": 0.58, "grad_norm": 0.4071083962917328, "learning_rate": 9.206762028608583e-05, "loss": 0.886, "step": 495 }, { "epoch": 0.58, "grad_norm": 0.4640657305717468, "learning_rate": 9.180754226267881e-05, "loss": 0.957, "step": 496 }, { "epoch": 0.59, "grad_norm": 0.4403044283390045, "learning_rate": 9.154746423927179e-05, "loss": 1.0355, "step": 497 }, { "epoch": 0.59, "grad_norm": 0.35679858922958374, "learning_rate": 9.128738621586477e-05, "loss": 0.9558, "step": 498 }, { "epoch": 0.59, "grad_norm": 0.3731861412525177, "learning_rate": 9.102730819245774e-05, "loss": 1.0631, "step": 499 }, { "epoch": 0.59, "grad_norm": 0.5025390386581421, "learning_rate": 9.076723016905072e-05, "loss": 0.9846, "step": 500 }, { "epoch": 0.59, "eval_loss": 0.9564683437347412, "eval_runtime": 545.8524, "eval_samples_per_second": 2.766, "eval_steps_per_second": 1.383, "step": 500 }, { "epoch": 0.59, "grad_norm": 0.3981812000274658, "learning_rate": 9.050715214564369e-05, "loss": 0.7666, "step": 501 }, { "epoch": 0.59, "grad_norm": 0.43981924653053284, "learning_rate": 9.024707412223667e-05, "loss": 0.8295, "step": 502 }, { "epoch": 0.59, "grad_norm": 0.9815853238105774, "learning_rate": 8.998699609882966e-05, "loss": 1.1342, "step": 503 }, { "epoch": 0.59, "grad_norm": 0.4028013050556183, "learning_rate": 8.972691807542262e-05, "loss": 0.9975, "step": 504 }, { "epoch": 0.59, "grad_norm": 0.9258779883384705, "learning_rate": 8.946684005201561e-05, "loss": 0.9955, "step": 505 }, { "epoch": 0.6, "grad_norm": 0.4801146388053894, "learning_rate": 8.920676202860858e-05, "loss": 1.0786, "step": 506 }, { "epoch": 0.6, "grad_norm": 0.4223402142524719, "learning_rate": 8.894668400520156e-05, "loss": 0.8817, "step": 507 }, { "epoch": 0.6, "grad_norm": 0.46179527044296265, "learning_rate": 8.868660598179454e-05, "loss": 0.8535, "step": 508 }, { "epoch": 0.6, "grad_norm": 0.7033700942993164, "learning_rate": 8.842652795838752e-05, "loss": 0.9261, "step": 509 }, { "epoch": 0.6, "grad_norm": 0.7166674733161926, "learning_rate": 8.81664499349805e-05, "loss": 1.0687, "step": 510 }, { "epoch": 0.6, "grad_norm": 0.6070614457130432, "learning_rate": 8.790637191157347e-05, "loss": 1.0352, "step": 511 }, { "epoch": 0.6, "grad_norm": 0.4146052598953247, "learning_rate": 8.764629388816646e-05, "loss": 0.8131, "step": 512 }, { "epoch": 0.6, "grad_norm": 0.45449090003967285, "learning_rate": 8.738621586475944e-05, "loss": 0.8424, "step": 513 }, { "epoch": 0.61, "grad_norm": 0.41242295503616333, "learning_rate": 8.71261378413524e-05, "loss": 0.8591, "step": 514 }, { "epoch": 0.61, "grad_norm": 0.4452085793018341, "learning_rate": 8.686605981794539e-05, "loss": 0.9126, "step": 515 }, { "epoch": 0.61, "grad_norm": 0.3656567633152008, "learning_rate": 8.660598179453837e-05, "loss": 0.7971, "step": 516 }, { "epoch": 0.61, "grad_norm": 0.5352630019187927, "learning_rate": 8.634590377113134e-05, "loss": 0.8963, "step": 517 }, { "epoch": 0.61, "grad_norm": 0.5082272291183472, "learning_rate": 8.608582574772432e-05, "loss": 0.9464, "step": 518 }, { "epoch": 0.61, "grad_norm": 0.47161418199539185, "learning_rate": 8.582574772431729e-05, "loss": 1.0905, "step": 519 }, { "epoch": 0.61, "grad_norm": 0.3810853660106659, "learning_rate": 8.556566970091027e-05, "loss": 0.8482, "step": 520 }, { "epoch": 0.61, "grad_norm": 0.37245139479637146, "learning_rate": 8.530559167750325e-05, "loss": 1.0261, "step": 521 }, { "epoch": 0.61, "grad_norm": 0.5589213967323303, "learning_rate": 8.504551365409624e-05, "loss": 0.8419, "step": 522 }, { "epoch": 0.62, "grad_norm": 0.34602364897727966, "learning_rate": 8.478543563068922e-05, "loss": 0.8963, "step": 523 }, { "epoch": 0.62, "grad_norm": 0.5899783372879028, "learning_rate": 8.452535760728219e-05, "loss": 0.7737, "step": 524 }, { "epoch": 0.62, "grad_norm": 0.40005919337272644, "learning_rate": 8.426527958387517e-05, "loss": 0.8447, "step": 525 }, { "epoch": 0.62, "grad_norm": 0.4106708765029907, "learning_rate": 8.400520156046815e-05, "loss": 0.9413, "step": 526 }, { "epoch": 0.62, "grad_norm": 0.28470200300216675, "learning_rate": 8.374512353706112e-05, "loss": 0.7514, "step": 527 }, { "epoch": 0.62, "grad_norm": 0.4649527370929718, "learning_rate": 8.34850455136541e-05, "loss": 0.9115, "step": 528 }, { "epoch": 0.62, "grad_norm": 0.4239949882030487, "learning_rate": 8.322496749024707e-05, "loss": 1.0008, "step": 529 }, { "epoch": 0.62, "grad_norm": 0.31371399760246277, "learning_rate": 8.296488946684005e-05, "loss": 0.7356, "step": 530 }, { "epoch": 0.63, "grad_norm": 0.5300579071044922, "learning_rate": 8.270481144343304e-05, "loss": 0.781, "step": 531 }, { "epoch": 0.63, "grad_norm": 0.4474256634712219, "learning_rate": 8.2444733420026e-05, "loss": 1.0941, "step": 532 }, { "epoch": 0.63, "grad_norm": 0.4346119463443756, "learning_rate": 8.218465539661899e-05, "loss": 0.7846, "step": 533 }, { "epoch": 0.63, "grad_norm": 0.626645028591156, "learning_rate": 8.192457737321197e-05, "loss": 0.9186, "step": 534 }, { "epoch": 0.63, "grad_norm": 0.4716324508190155, "learning_rate": 8.166449934980495e-05, "loss": 1.0241, "step": 535 }, { "epoch": 0.63, "grad_norm": 0.41494086384773254, "learning_rate": 8.140442132639793e-05, "loss": 0.8293, "step": 536 }, { "epoch": 0.63, "grad_norm": 0.35419961810112, "learning_rate": 8.11443433029909e-05, "loss": 1.2024, "step": 537 }, { "epoch": 0.63, "grad_norm": 0.3790522813796997, "learning_rate": 8.088426527958388e-05, "loss": 0.9508, "step": 538 }, { "epoch": 0.63, "grad_norm": 0.6440338492393494, "learning_rate": 8.062418725617685e-05, "loss": 1.0345, "step": 539 }, { "epoch": 0.64, "grad_norm": 0.5094674825668335, "learning_rate": 8.036410923276983e-05, "loss": 1.0392, "step": 540 }, { "epoch": 0.64, "grad_norm": 0.38184013962745667, "learning_rate": 8.010403120936282e-05, "loss": 1.0548, "step": 541 }, { "epoch": 0.64, "grad_norm": 0.5326213240623474, "learning_rate": 7.984395318595579e-05, "loss": 0.924, "step": 542 }, { "epoch": 0.64, "grad_norm": 0.6736549735069275, "learning_rate": 7.958387516254877e-05, "loss": 0.9433, "step": 543 }, { "epoch": 0.64, "grad_norm": 0.45961499214172363, "learning_rate": 7.932379713914174e-05, "loss": 0.8909, "step": 544 }, { "epoch": 0.64, "grad_norm": 0.6285248398780823, "learning_rate": 7.906371911573472e-05, "loss": 0.6791, "step": 545 }, { "epoch": 0.64, "grad_norm": 0.35338294506073, "learning_rate": 7.88036410923277e-05, "loss": 0.9138, "step": 546 }, { "epoch": 0.64, "grad_norm": 0.6357168555259705, "learning_rate": 7.854356306892068e-05, "loss": 0.8857, "step": 547 }, { "epoch": 0.65, "grad_norm": 0.5306888818740845, "learning_rate": 7.828348504551366e-05, "loss": 0.9735, "step": 548 }, { "epoch": 0.65, "grad_norm": 0.4110960364341736, "learning_rate": 7.802340702210663e-05, "loss": 0.8371, "step": 549 }, { "epoch": 0.65, "grad_norm": 0.521158754825592, "learning_rate": 7.776332899869962e-05, "loss": 1.0281, "step": 550 }, { "epoch": 0.65, "grad_norm": 0.4075251817703247, "learning_rate": 7.75032509752926e-05, "loss": 0.9324, "step": 551 }, { "epoch": 0.65, "grad_norm": 0.3983442783355713, "learning_rate": 7.724317295188557e-05, "loss": 1.0884, "step": 552 }, { "epoch": 0.65, "grad_norm": 0.45374634861946106, "learning_rate": 7.698309492847855e-05, "loss": 0.9971, "step": 553 }, { "epoch": 0.65, "grad_norm": 0.4778593182563782, "learning_rate": 7.672301690507152e-05, "loss": 1.1167, "step": 554 }, { "epoch": 0.65, "grad_norm": 0.42328914999961853, "learning_rate": 7.64629388816645e-05, "loss": 1.0628, "step": 555 }, { "epoch": 0.65, "grad_norm": 0.3874073326587677, "learning_rate": 7.620286085825748e-05, "loss": 0.8475, "step": 556 }, { "epoch": 0.66, "grad_norm": 0.3884342610836029, "learning_rate": 7.594278283485045e-05, "loss": 1.0887, "step": 557 }, { "epoch": 0.66, "grad_norm": 0.6277136206626892, "learning_rate": 7.568270481144345e-05, "loss": 0.8782, "step": 558 }, { "epoch": 0.66, "grad_norm": 0.5219310522079468, "learning_rate": 7.542262678803641e-05, "loss": 1.0388, "step": 559 }, { "epoch": 0.66, "grad_norm": 0.3901873528957367, "learning_rate": 7.51625487646294e-05, "loss": 1.0146, "step": 560 }, { "epoch": 0.66, "grad_norm": 0.4146654009819031, "learning_rate": 7.490247074122238e-05, "loss": 0.9669, "step": 561 }, { "epoch": 0.66, "grad_norm": 0.35999563336372375, "learning_rate": 7.464239271781535e-05, "loss": 1.017, "step": 562 }, { "epoch": 0.66, "grad_norm": 0.42414629459381104, "learning_rate": 7.438231469440833e-05, "loss": 1.0244, "step": 563 }, { "epoch": 0.66, "grad_norm": 0.4917350113391876, "learning_rate": 7.41222366710013e-05, "loss": 1.0007, "step": 564 }, { "epoch": 0.67, "grad_norm": 0.44585248827934265, "learning_rate": 7.386215864759428e-05, "loss": 0.8734, "step": 565 }, { "epoch": 0.67, "grad_norm": 0.4269184470176697, "learning_rate": 7.360208062418726e-05, "loss": 0.8579, "step": 566 }, { "epoch": 0.67, "grad_norm": 0.36958837509155273, "learning_rate": 7.334200260078023e-05, "loss": 0.6676, "step": 567 }, { "epoch": 0.67, "grad_norm": 0.3978622555732727, "learning_rate": 7.308192457737321e-05, "loss": 0.9116, "step": 568 }, { "epoch": 0.67, "grad_norm": 0.349263459444046, "learning_rate": 7.28218465539662e-05, "loss": 0.8394, "step": 569 }, { "epoch": 0.67, "grad_norm": 0.3653329014778137, "learning_rate": 7.256176853055918e-05, "loss": 0.8487, "step": 570 }, { "epoch": 0.67, "grad_norm": 0.40273159742355347, "learning_rate": 7.230169050715216e-05, "loss": 0.9374, "step": 571 }, { "epoch": 0.67, "grad_norm": 0.40544393658638, "learning_rate": 7.204161248374513e-05, "loss": 0.6764, "step": 572 }, { "epoch": 0.67, "grad_norm": 0.38676464557647705, "learning_rate": 7.178153446033811e-05, "loss": 0.9505, "step": 573 }, { "epoch": 0.68, "grad_norm": 0.5303266048431396, "learning_rate": 7.152145643693108e-05, "loss": 0.9271, "step": 574 }, { "epoch": 0.68, "grad_norm": 0.42933189868927, "learning_rate": 7.126137841352406e-05, "loss": 0.8537, "step": 575 }, { "epoch": 0.68, "grad_norm": 0.3899463415145874, "learning_rate": 7.100130039011704e-05, "loss": 0.915, "step": 576 }, { "epoch": 0.68, "grad_norm": 0.4377955496311188, "learning_rate": 7.074122236671001e-05, "loss": 0.7395, "step": 577 }, { "epoch": 0.68, "grad_norm": 0.4030522108078003, "learning_rate": 7.0481144343303e-05, "loss": 1.0386, "step": 578 }, { "epoch": 0.68, "grad_norm": 0.5526483058929443, "learning_rate": 7.022106631989596e-05, "loss": 0.9547, "step": 579 }, { "epoch": 0.68, "grad_norm": 0.7406371235847473, "learning_rate": 6.996098829648895e-05, "loss": 0.981, "step": 580 }, { "epoch": 0.68, "grad_norm": 0.584095299243927, "learning_rate": 6.970091027308193e-05, "loss": 0.7351, "step": 581 }, { "epoch": 0.69, "grad_norm": 0.4012065529823303, "learning_rate": 6.944083224967491e-05, "loss": 0.8546, "step": 582 }, { "epoch": 0.69, "grad_norm": 0.45636311173439026, "learning_rate": 6.918075422626789e-05, "loss": 1.1251, "step": 583 }, { "epoch": 0.69, "grad_norm": 0.4690900146961212, "learning_rate": 6.892067620286086e-05, "loss": 0.9391, "step": 584 }, { "epoch": 0.69, "grad_norm": 0.3936656415462494, "learning_rate": 6.866059817945384e-05, "loss": 0.8532, "step": 585 }, { "epoch": 0.69, "grad_norm": 0.3015373647212982, "learning_rate": 6.840052015604682e-05, "loss": 0.7941, "step": 586 }, { "epoch": 0.69, "grad_norm": 0.4804030954837799, "learning_rate": 6.81404421326398e-05, "loss": 0.8089, "step": 587 }, { "epoch": 0.69, "grad_norm": 0.6172294020652771, "learning_rate": 6.788036410923278e-05, "loss": 1.1117, "step": 588 }, { "epoch": 0.69, "grad_norm": 0.4950728416442871, "learning_rate": 6.762028608582574e-05, "loss": 0.9228, "step": 589 }, { "epoch": 0.69, "grad_norm": 0.4423811137676239, "learning_rate": 6.736020806241873e-05, "loss": 0.9665, "step": 590 }, { "epoch": 0.7, "grad_norm": 0.5107194781303406, "learning_rate": 6.710013003901171e-05, "loss": 0.9299, "step": 591 }, { "epoch": 0.7, "grad_norm": 0.6242122054100037, "learning_rate": 6.684005201560468e-05, "loss": 0.8876, "step": 592 }, { "epoch": 0.7, "grad_norm": 0.3378008306026459, "learning_rate": 6.657997399219766e-05, "loss": 0.9089, "step": 593 }, { "epoch": 0.7, "grad_norm": 0.3389635980129242, "learning_rate": 6.631989596879064e-05, "loss": 0.7632, "step": 594 }, { "epoch": 0.7, "grad_norm": 0.38348597288131714, "learning_rate": 6.605981794538362e-05, "loss": 1.1742, "step": 595 }, { "epoch": 0.7, "grad_norm": 0.40138015151023865, "learning_rate": 6.57997399219766e-05, "loss": 1.0699, "step": 596 }, { "epoch": 0.7, "grad_norm": 0.4016960859298706, "learning_rate": 6.553966189856957e-05, "loss": 0.7236, "step": 597 }, { "epoch": 0.7, "grad_norm": 0.5351032614707947, "learning_rate": 6.527958387516256e-05, "loss": 0.9702, "step": 598 }, { "epoch": 0.71, "grad_norm": 0.4079533517360687, "learning_rate": 6.501950585175553e-05, "loss": 0.7466, "step": 599 }, { "epoch": 0.71, "grad_norm": 0.3392191529273987, "learning_rate": 6.475942782834851e-05, "loss": 0.9523, "step": 600 }, { "epoch": 0.71, "eval_loss": 0.9486369490623474, "eval_runtime": 546.4506, "eval_samples_per_second": 2.763, "eval_steps_per_second": 1.382, "step": 600 }, { "epoch": 0.71, "grad_norm": 0.5096669793128967, "learning_rate": 6.449934980494149e-05, "loss": 0.9551, "step": 601 }, { "epoch": 0.71, "grad_norm": 0.4873715341091156, "learning_rate": 6.423927178153446e-05, "loss": 0.8347, "step": 602 }, { "epoch": 0.71, "grad_norm": 0.386218398809433, "learning_rate": 6.397919375812744e-05, "loss": 0.9135, "step": 603 }, { "epoch": 0.71, "grad_norm": 0.6417282223701477, "learning_rate": 6.371911573472041e-05, "loss": 0.8427, "step": 604 }, { "epoch": 0.71, "grad_norm": 0.4092877209186554, "learning_rate": 6.345903771131339e-05, "loss": 0.982, "step": 605 }, { "epoch": 0.71, "grad_norm": 0.3618543744087219, "learning_rate": 6.319895968790637e-05, "loss": 0.9055, "step": 606 }, { "epoch": 0.71, "grad_norm": 0.40482065081596375, "learning_rate": 6.293888166449936e-05, "loss": 0.9119, "step": 607 }, { "epoch": 0.72, "grad_norm": 0.7335551381111145, "learning_rate": 6.267880364109234e-05, "loss": 0.9509, "step": 608 }, { "epoch": 0.72, "grad_norm": 0.5208268761634827, "learning_rate": 6.241872561768532e-05, "loss": 0.8131, "step": 609 }, { "epoch": 0.72, "grad_norm": 0.5012003779411316, "learning_rate": 6.215864759427829e-05, "loss": 0.8131, "step": 610 }, { "epoch": 0.72, "grad_norm": 0.3858044147491455, "learning_rate": 6.189856957087127e-05, "loss": 1.026, "step": 611 }, { "epoch": 0.72, "grad_norm": 0.5440144538879395, "learning_rate": 6.163849154746424e-05, "loss": 0.874, "step": 612 }, { "epoch": 0.72, "grad_norm": 0.37393397092819214, "learning_rate": 6.137841352405722e-05, "loss": 0.9728, "step": 613 }, { "epoch": 0.72, "grad_norm": 0.4478467106819153, "learning_rate": 6.11183355006502e-05, "loss": 0.9412, "step": 614 }, { "epoch": 0.72, "grad_norm": 0.5505075454711914, "learning_rate": 6.085825747724317e-05, "loss": 0.947, "step": 615 }, { "epoch": 0.73, "grad_norm": 0.3153130114078522, "learning_rate": 6.0598179453836155e-05, "loss": 0.8113, "step": 616 }, { "epoch": 0.73, "grad_norm": 0.5335275530815125, "learning_rate": 6.033810143042913e-05, "loss": 0.7609, "step": 617 }, { "epoch": 0.73, "grad_norm": 0.6180243492126465, "learning_rate": 6.007802340702211e-05, "loss": 0.8649, "step": 618 }, { "epoch": 0.73, "grad_norm": 0.4858168363571167, "learning_rate": 5.9817945383615094e-05, "loss": 0.8953, "step": 619 }, { "epoch": 0.73, "grad_norm": 0.36638808250427246, "learning_rate": 5.955786736020806e-05, "loss": 1.0415, "step": 620 }, { "epoch": 0.73, "grad_norm": 0.41763070225715637, "learning_rate": 5.9297789336801045e-05, "loss": 1.0474, "step": 621 }, { "epoch": 0.73, "grad_norm": 0.6354649662971497, "learning_rate": 5.903771131339402e-05, "loss": 1.0198, "step": 622 }, { "epoch": 0.73, "grad_norm": 0.38064074516296387, "learning_rate": 5.8777633289987e-05, "loss": 0.9917, "step": 623 }, { "epoch": 0.73, "grad_norm": 0.5428590774536133, "learning_rate": 5.8517555266579985e-05, "loss": 0.8644, "step": 624 }, { "epoch": 0.74, "grad_norm": 0.39888277649879456, "learning_rate": 5.8257477243172954e-05, "loss": 0.9641, "step": 625 }, { "epoch": 0.74, "grad_norm": 0.35818690061569214, "learning_rate": 5.7997399219765936e-05, "loss": 0.958, "step": 626 }, { "epoch": 0.74, "grad_norm": 0.39909327030181885, "learning_rate": 5.7737321196358904e-05, "loss": 0.9675, "step": 627 }, { "epoch": 0.74, "grad_norm": 0.48225390911102295, "learning_rate": 5.747724317295189e-05, "loss": 0.8243, "step": 628 }, { "epoch": 0.74, "grad_norm": 0.5696926116943359, "learning_rate": 5.721716514954487e-05, "loss": 0.7682, "step": 629 }, { "epoch": 0.74, "grad_norm": 0.4227491021156311, "learning_rate": 5.6957087126137844e-05, "loss": 1.1601, "step": 630 }, { "epoch": 0.74, "grad_norm": 0.43947187066078186, "learning_rate": 5.6697009102730826e-05, "loss": 0.8134, "step": 631 }, { "epoch": 0.74, "grad_norm": 0.5529877543449402, "learning_rate": 5.6436931079323795e-05, "loss": 1.1165, "step": 632 }, { "epoch": 0.75, "grad_norm": 0.45177382230758667, "learning_rate": 5.617685305591678e-05, "loss": 0.9771, "step": 633 }, { "epoch": 0.75, "grad_norm": 0.5075746774673462, "learning_rate": 5.591677503250976e-05, "loss": 0.7307, "step": 634 }, { "epoch": 0.75, "grad_norm": 0.518578290939331, "learning_rate": 5.5656697009102735e-05, "loss": 0.9702, "step": 635 }, { "epoch": 0.75, "grad_norm": 0.2868906259536743, "learning_rate": 5.539661898569572e-05, "loss": 0.9572, "step": 636 }, { "epoch": 0.75, "grad_norm": 0.4592292010784149, "learning_rate": 5.5136540962288686e-05, "loss": 0.7495, "step": 637 }, { "epoch": 0.75, "grad_norm": 0.442264586687088, "learning_rate": 5.487646293888167e-05, "loss": 0.9416, "step": 638 }, { "epoch": 0.75, "grad_norm": 0.3862535059452057, "learning_rate": 5.461638491547465e-05, "loss": 1.0871, "step": 639 }, { "epoch": 0.75, "grad_norm": 0.3567231297492981, "learning_rate": 5.435630689206762e-05, "loss": 0.9585, "step": 640 }, { "epoch": 0.76, "grad_norm": 0.35670965909957886, "learning_rate": 5.40962288686606e-05, "loss": 0.8974, "step": 641 }, { "epoch": 0.76, "grad_norm": 0.33253470063209534, "learning_rate": 5.3836150845253576e-05, "loss": 0.7088, "step": 642 }, { "epoch": 0.76, "grad_norm": 0.545925498008728, "learning_rate": 5.357607282184656e-05, "loss": 0.9076, "step": 643 }, { "epoch": 0.76, "grad_norm": 0.3469606935977936, "learning_rate": 5.331599479843954e-05, "loss": 0.6942, "step": 644 }, { "epoch": 0.76, "grad_norm": 0.645745038986206, "learning_rate": 5.305591677503251e-05, "loss": 0.845, "step": 645 }, { "epoch": 0.76, "grad_norm": 0.44218504428863525, "learning_rate": 5.279583875162549e-05, "loss": 1.0942, "step": 646 }, { "epoch": 0.76, "grad_norm": 0.40422385931015015, "learning_rate": 5.253576072821847e-05, "loss": 1.1837, "step": 647 }, { "epoch": 0.76, "grad_norm": 0.3621053695678711, "learning_rate": 5.227568270481145e-05, "loss": 0.9409, "step": 648 }, { "epoch": 0.76, "grad_norm": 0.5237849950790405, "learning_rate": 5.201560468140443e-05, "loss": 0.9584, "step": 649 }, { "epoch": 0.77, "grad_norm": 0.42268675565719604, "learning_rate": 5.17555266579974e-05, "loss": 0.918, "step": 650 }, { "epoch": 0.77, "grad_norm": 0.6781652569770813, "learning_rate": 5.149544863459038e-05, "loss": 0.8015, "step": 651 }, { "epoch": 0.77, "grad_norm": 0.4573494493961334, "learning_rate": 5.123537061118335e-05, "loss": 0.8392, "step": 652 }, { "epoch": 0.77, "grad_norm": 0.48572808504104614, "learning_rate": 5.097529258777633e-05, "loss": 0.833, "step": 653 }, { "epoch": 0.77, "grad_norm": 0.4242917597293854, "learning_rate": 5.0715214564369315e-05, "loss": 0.6915, "step": 654 }, { "epoch": 0.77, "grad_norm": 0.4147329032421112, "learning_rate": 5.045513654096229e-05, "loss": 0.8213, "step": 655 }, { "epoch": 0.77, "grad_norm": 0.6368014216423035, "learning_rate": 5.019505851755527e-05, "loss": 0.9627, "step": 656 }, { "epoch": 0.77, "grad_norm": 0.44350188970565796, "learning_rate": 4.993498049414825e-05, "loss": 0.8598, "step": 657 }, { "epoch": 0.78, "grad_norm": 0.5331006050109863, "learning_rate": 4.967490247074122e-05, "loss": 0.8663, "step": 658 }, { "epoch": 0.78, "grad_norm": 0.498432457447052, "learning_rate": 4.94148244473342e-05, "loss": 0.9025, "step": 659 }, { "epoch": 0.78, "grad_norm": 0.4427286684513092, "learning_rate": 4.915474642392718e-05, "loss": 0.9307, "step": 660 }, { "epoch": 0.78, "grad_norm": 0.3966609537601471, "learning_rate": 4.889466840052016e-05, "loss": 0.971, "step": 661 }, { "epoch": 0.78, "grad_norm": 0.582984209060669, "learning_rate": 4.863459037711314e-05, "loss": 1.1528, "step": 662 }, { "epoch": 0.78, "grad_norm": 0.40987858176231384, "learning_rate": 4.8374512353706114e-05, "loss": 0.8503, "step": 663 }, { "epoch": 0.78, "grad_norm": 0.4146847426891327, "learning_rate": 4.811443433029909e-05, "loss": 0.8022, "step": 664 }, { "epoch": 0.78, "grad_norm": 0.5389853119850159, "learning_rate": 4.7854356306892065e-05, "loss": 1.0602, "step": 665 }, { "epoch": 0.78, "grad_norm": 0.4369354248046875, "learning_rate": 4.759427828348505e-05, "loss": 0.9696, "step": 666 }, { "epoch": 0.79, "grad_norm": 0.35021817684173584, "learning_rate": 4.733420026007803e-05, "loss": 0.8269, "step": 667 }, { "epoch": 0.79, "grad_norm": 0.5168956518173218, "learning_rate": 4.7074122236671004e-05, "loss": 0.903, "step": 668 }, { "epoch": 0.79, "grad_norm": 0.39560142159461975, "learning_rate": 4.681404421326398e-05, "loss": 0.6843, "step": 669 }, { "epoch": 0.79, "grad_norm": 0.31525564193725586, "learning_rate": 4.6553966189856955e-05, "loss": 0.9849, "step": 670 }, { "epoch": 0.79, "grad_norm": 0.762050986289978, "learning_rate": 4.629388816644994e-05, "loss": 0.8766, "step": 671 }, { "epoch": 0.79, "grad_norm": 0.3123376667499542, "learning_rate": 4.603381014304291e-05, "loss": 0.9056, "step": 672 }, { "epoch": 0.79, "grad_norm": 0.39626839756965637, "learning_rate": 4.5773732119635895e-05, "loss": 0.9783, "step": 673 }, { "epoch": 0.79, "grad_norm": 0.5586526989936829, "learning_rate": 4.551365409622887e-05, "loss": 1.1036, "step": 674 }, { "epoch": 0.8, "grad_norm": 0.5380015969276428, "learning_rate": 4.5253576072821846e-05, "loss": 0.8872, "step": 675 }, { "epoch": 0.8, "grad_norm": 0.46103811264038086, "learning_rate": 4.499349804941483e-05, "loss": 0.7392, "step": 676 }, { "epoch": 0.8, "grad_norm": 0.34783852100372314, "learning_rate": 4.4733420026007803e-05, "loss": 0.8752, "step": 677 }, { "epoch": 0.8, "grad_norm": 0.3441290855407715, "learning_rate": 4.447334200260078e-05, "loss": 0.7894, "step": 678 }, { "epoch": 0.8, "grad_norm": 0.6663783192634583, "learning_rate": 4.421326397919376e-05, "loss": 1.0241, "step": 679 }, { "epoch": 0.8, "grad_norm": 0.43966543674468994, "learning_rate": 4.3953185955786736e-05, "loss": 0.5997, "step": 680 }, { "epoch": 0.8, "grad_norm": 0.6549209356307983, "learning_rate": 4.369310793237972e-05, "loss": 0.9197, "step": 681 }, { "epoch": 0.8, "grad_norm": 0.532124400138855, "learning_rate": 4.3433029908972694e-05, "loss": 1.0531, "step": 682 }, { "epoch": 0.8, "grad_norm": 0.3629736304283142, "learning_rate": 4.317295188556567e-05, "loss": 0.8537, "step": 683 }, { "epoch": 0.81, "grad_norm": 0.30578306317329407, "learning_rate": 4.2912873862158645e-05, "loss": 0.7, "step": 684 }, { "epoch": 0.81, "grad_norm": 0.40250301361083984, "learning_rate": 4.265279583875163e-05, "loss": 0.8544, "step": 685 }, { "epoch": 0.81, "grad_norm": 0.6847468614578247, "learning_rate": 4.239271781534461e-05, "loss": 1.1026, "step": 686 }, { "epoch": 0.81, "grad_norm": 0.5002169609069824, "learning_rate": 4.2132639791937585e-05, "loss": 0.6336, "step": 687 }, { "epoch": 0.81, "grad_norm": 0.4745190739631653, "learning_rate": 4.187256176853056e-05, "loss": 0.8971, "step": 688 }, { "epoch": 0.81, "grad_norm": 0.7550213932991028, "learning_rate": 4.1612483745123535e-05, "loss": 1.1197, "step": 689 }, { "epoch": 0.81, "grad_norm": 0.4807836711406708, "learning_rate": 4.135240572171652e-05, "loss": 0.8011, "step": 690 }, { "epoch": 0.81, "grad_norm": 0.5241039991378784, "learning_rate": 4.109232769830949e-05, "loss": 0.8434, "step": 691 }, { "epoch": 0.82, "grad_norm": 0.4142901599407196, "learning_rate": 4.0832249674902475e-05, "loss": 0.8922, "step": 692 }, { "epoch": 0.82, "grad_norm": 0.5607759952545166, "learning_rate": 4.057217165149545e-05, "loss": 1.0014, "step": 693 }, { "epoch": 0.82, "grad_norm": 0.45777449011802673, "learning_rate": 4.0312093628088426e-05, "loss": 0.975, "step": 694 }, { "epoch": 0.82, "grad_norm": 0.5091066360473633, "learning_rate": 4.005201560468141e-05, "loss": 0.9695, "step": 695 }, { "epoch": 0.82, "grad_norm": 0.4820137619972229, "learning_rate": 3.9791937581274384e-05, "loss": 0.9444, "step": 696 }, { "epoch": 0.82, "grad_norm": 0.3295759856700897, "learning_rate": 3.953185955786736e-05, "loss": 0.8029, "step": 697 }, { "epoch": 0.82, "grad_norm": 0.3466610014438629, "learning_rate": 3.927178153446034e-05, "loss": 0.8585, "step": 698 }, { "epoch": 0.82, "grad_norm": 0.4017598032951355, "learning_rate": 3.9011703511053317e-05, "loss": 0.9429, "step": 699 }, { "epoch": 0.82, "grad_norm": 0.45068544149398804, "learning_rate": 3.87516254876463e-05, "loss": 0.8567, "step": 700 }, { "epoch": 0.82, "eval_loss": 0.9429731369018555, "eval_runtime": 545.935, "eval_samples_per_second": 2.766, "eval_steps_per_second": 1.383, "step": 700 }, { "epoch": 0.83, "grad_norm": 0.3644198179244995, "learning_rate": 3.8491547464239274e-05, "loss": 0.8288, "step": 701 }, { "epoch": 0.83, "grad_norm": 0.46298202872276306, "learning_rate": 3.823146944083225e-05, "loss": 0.9233, "step": 702 }, { "epoch": 0.83, "grad_norm": 0.38276007771492004, "learning_rate": 3.7971391417425225e-05, "loss": 0.8703, "step": 703 }, { "epoch": 0.83, "grad_norm": 0.3996793031692505, "learning_rate": 3.771131339401821e-05, "loss": 1.0333, "step": 704 }, { "epoch": 0.83, "grad_norm": 0.4469836950302124, "learning_rate": 3.745123537061119e-05, "loss": 0.9521, "step": 705 }, { "epoch": 0.83, "grad_norm": 0.42836982011795044, "learning_rate": 3.7191157347204165e-05, "loss": 0.6851, "step": 706 }, { "epoch": 0.83, "grad_norm": 0.38192087411880493, "learning_rate": 3.693107932379714e-05, "loss": 0.9778, "step": 707 }, { "epoch": 0.83, "grad_norm": 0.3983215391635895, "learning_rate": 3.6671001300390116e-05, "loss": 0.9587, "step": 708 }, { "epoch": 0.84, "grad_norm": 0.38901060819625854, "learning_rate": 3.64109232769831e-05, "loss": 0.7352, "step": 709 }, { "epoch": 0.84, "grad_norm": 0.4101078510284424, "learning_rate": 3.615084525357608e-05, "loss": 0.813, "step": 710 }, { "epoch": 0.84, "grad_norm": 0.41320496797561646, "learning_rate": 3.5890767230169055e-05, "loss": 0.9269, "step": 711 }, { "epoch": 0.84, "grad_norm": 0.35738593339920044, "learning_rate": 3.563068920676203e-05, "loss": 1.0962, "step": 712 }, { "epoch": 0.84, "grad_norm": 0.6265596151351929, "learning_rate": 3.5370611183355006e-05, "loss": 0.8924, "step": 713 }, { "epoch": 0.84, "grad_norm": 0.6120777130126953, "learning_rate": 3.511053315994798e-05, "loss": 0.865, "step": 714 }, { "epoch": 0.84, "grad_norm": 0.36191537976264954, "learning_rate": 3.4850455136540964e-05, "loss": 0.8703, "step": 715 }, { "epoch": 0.84, "grad_norm": 0.4391356408596039, "learning_rate": 3.4590377113133946e-05, "loss": 1.0831, "step": 716 }, { "epoch": 0.84, "grad_norm": 0.5841166377067566, "learning_rate": 3.433029908972692e-05, "loss": 0.8891, "step": 717 }, { "epoch": 0.85, "grad_norm": 0.4388696551322937, "learning_rate": 3.40702210663199e-05, "loss": 0.9449, "step": 718 }, { "epoch": 0.85, "grad_norm": 0.3359023332595825, "learning_rate": 3.381014304291287e-05, "loss": 0.7529, "step": 719 }, { "epoch": 0.85, "grad_norm": 0.4563657343387604, "learning_rate": 3.3550065019505854e-05, "loss": 1.0695, "step": 720 }, { "epoch": 0.85, "grad_norm": 0.4048353433609009, "learning_rate": 3.328998699609883e-05, "loss": 0.933, "step": 721 }, { "epoch": 0.85, "grad_norm": 0.4353731870651245, "learning_rate": 3.302990897269181e-05, "loss": 0.9173, "step": 722 }, { "epoch": 0.85, "grad_norm": 0.3326166570186615, "learning_rate": 3.276983094928479e-05, "loss": 1.064, "step": 723 }, { "epoch": 0.85, "grad_norm": 0.6150364279747009, "learning_rate": 3.250975292587776e-05, "loss": 0.9995, "step": 724 }, { "epoch": 0.85, "grad_norm": 0.5298452973365784, "learning_rate": 3.2249674902470745e-05, "loss": 1.0936, "step": 725 }, { "epoch": 0.86, "grad_norm": 0.4979341924190521, "learning_rate": 3.198959687906372e-05, "loss": 0.6313, "step": 726 }, { "epoch": 0.86, "grad_norm": 0.4423145055770874, "learning_rate": 3.1729518855656696e-05, "loss": 0.7411, "step": 727 }, { "epoch": 0.86, "grad_norm": 0.46717703342437744, "learning_rate": 3.146944083224968e-05, "loss": 0.8921, "step": 728 }, { "epoch": 0.86, "grad_norm": 0.6488456726074219, "learning_rate": 3.120936280884266e-05, "loss": 0.9583, "step": 729 }, { "epoch": 0.86, "grad_norm": 0.44180455803871155, "learning_rate": 3.0949284785435635e-05, "loss": 0.8997, "step": 730 }, { "epoch": 0.86, "grad_norm": 0.47081491351127625, "learning_rate": 3.068920676202861e-05, "loss": 0.7978, "step": 731 }, { "epoch": 0.86, "grad_norm": 0.7937029600143433, "learning_rate": 3.0429128738621586e-05, "loss": 1.0736, "step": 732 }, { "epoch": 0.86, "grad_norm": 0.6148236393928528, "learning_rate": 3.0169050715214565e-05, "loss": 0.8235, "step": 733 }, { "epoch": 0.86, "grad_norm": 0.4343833029270172, "learning_rate": 2.9908972691807547e-05, "loss": 0.8299, "step": 734 }, { "epoch": 0.87, "grad_norm": 0.4517049193382263, "learning_rate": 2.9648894668400523e-05, "loss": 1.1623, "step": 735 }, { "epoch": 0.87, "grad_norm": 0.5517309308052063, "learning_rate": 2.93888166449935e-05, "loss": 1.1099, "step": 736 }, { "epoch": 0.87, "grad_norm": 0.7555133700370789, "learning_rate": 2.9128738621586477e-05, "loss": 1.0696, "step": 737 }, { "epoch": 0.87, "grad_norm": 0.4421709179878235, "learning_rate": 2.8868660598179452e-05, "loss": 0.9107, "step": 738 }, { "epoch": 0.87, "grad_norm": 0.379757821559906, "learning_rate": 2.8608582574772434e-05, "loss": 0.7525, "step": 739 }, { "epoch": 0.87, "grad_norm": 0.4668852984905243, "learning_rate": 2.8348504551365413e-05, "loss": 0.8708, "step": 740 }, { "epoch": 0.87, "grad_norm": 0.3945755362510681, "learning_rate": 2.808842652795839e-05, "loss": 0.8768, "step": 741 }, { "epoch": 0.87, "grad_norm": 0.5074703693389893, "learning_rate": 2.7828348504551367e-05, "loss": 0.8388, "step": 742 }, { "epoch": 0.88, "grad_norm": 0.426055371761322, "learning_rate": 2.7568270481144343e-05, "loss": 1.0121, "step": 743 }, { "epoch": 0.88, "grad_norm": 0.6159988045692444, "learning_rate": 2.7308192457737325e-05, "loss": 0.8153, "step": 744 }, { "epoch": 0.88, "grad_norm": 0.4640282094478607, "learning_rate": 2.70481144343303e-05, "loss": 0.9953, "step": 745 }, { "epoch": 0.88, "grad_norm": 0.3854639232158661, "learning_rate": 2.678803641092328e-05, "loss": 0.9826, "step": 746 }, { "epoch": 0.88, "grad_norm": 0.4375513195991516, "learning_rate": 2.6527958387516255e-05, "loss": 1.1403, "step": 747 }, { "epoch": 0.88, "grad_norm": 0.5711820125579834, "learning_rate": 2.6267880364109233e-05, "loss": 0.7825, "step": 748 }, { "epoch": 0.88, "grad_norm": 0.8451410531997681, "learning_rate": 2.6007802340702216e-05, "loss": 1.0823, "step": 749 }, { "epoch": 0.88, "grad_norm": 0.47502952814102173, "learning_rate": 2.574772431729519e-05, "loss": 0.8728, "step": 750 }, { "epoch": 0.88, "grad_norm": 0.4285711348056793, "learning_rate": 2.5487646293888166e-05, "loss": 1.0179, "step": 751 }, { "epoch": 0.89, "grad_norm": 0.8021396398544312, "learning_rate": 2.5227568270481145e-05, "loss": 1.0363, "step": 752 }, { "epoch": 0.89, "grad_norm": 0.7019473910331726, "learning_rate": 2.4967490247074124e-05, "loss": 0.9126, "step": 753 }, { "epoch": 0.89, "grad_norm": 0.3320355713367462, "learning_rate": 2.47074122236671e-05, "loss": 0.6946, "step": 754 }, { "epoch": 0.89, "grad_norm": 0.4108569622039795, "learning_rate": 2.444733420026008e-05, "loss": 0.8639, "step": 755 }, { "epoch": 0.89, "grad_norm": 0.34045693278312683, "learning_rate": 2.4187256176853057e-05, "loss": 0.8135, "step": 756 }, { "epoch": 0.89, "grad_norm": 0.5254625082015991, "learning_rate": 2.3927178153446032e-05, "loss": 0.8942, "step": 757 }, { "epoch": 0.89, "grad_norm": 0.41317030787467957, "learning_rate": 2.3667100130039015e-05, "loss": 0.9522, "step": 758 }, { "epoch": 0.89, "grad_norm": 0.7283076643943787, "learning_rate": 2.340702210663199e-05, "loss": 0.977, "step": 759 }, { "epoch": 0.9, "grad_norm": 0.45741039514541626, "learning_rate": 2.314694408322497e-05, "loss": 0.7947, "step": 760 }, { "epoch": 0.9, "grad_norm": 0.4744103252887726, "learning_rate": 2.2886866059817948e-05, "loss": 0.8825, "step": 761 }, { "epoch": 0.9, "grad_norm": 0.4509696364402771, "learning_rate": 2.2626788036410923e-05, "loss": 0.9413, "step": 762 }, { "epoch": 0.9, "grad_norm": 0.4502064883708954, "learning_rate": 2.2366710013003902e-05, "loss": 0.7723, "step": 763 }, { "epoch": 0.9, "grad_norm": 0.5345553755760193, "learning_rate": 2.210663198959688e-05, "loss": 1.0332, "step": 764 }, { "epoch": 0.9, "grad_norm": 0.7038628458976746, "learning_rate": 2.184655396618986e-05, "loss": 0.7661, "step": 765 }, { "epoch": 0.9, "grad_norm": 0.48555681109428406, "learning_rate": 2.1586475942782835e-05, "loss": 0.9913, "step": 766 }, { "epoch": 0.9, "grad_norm": 0.4042430818080902, "learning_rate": 2.1326397919375814e-05, "loss": 1.0808, "step": 767 }, { "epoch": 0.9, "grad_norm": 0.42962175607681274, "learning_rate": 2.1066319895968792e-05, "loss": 0.8371, "step": 768 }, { "epoch": 0.91, "grad_norm": 0.49653828144073486, "learning_rate": 2.0806241872561768e-05, "loss": 0.946, "step": 769 }, { "epoch": 0.91, "grad_norm": 0.44352245330810547, "learning_rate": 2.0546163849154747e-05, "loss": 0.8473, "step": 770 }, { "epoch": 0.91, "grad_norm": 0.46388956904411316, "learning_rate": 2.0286085825747725e-05, "loss": 1.0095, "step": 771 }, { "epoch": 0.91, "grad_norm": 0.6533496379852295, "learning_rate": 2.0026007802340704e-05, "loss": 0.8451, "step": 772 }, { "epoch": 0.91, "grad_norm": 0.4906177222728729, "learning_rate": 1.976592977893368e-05, "loss": 1.0586, "step": 773 }, { "epoch": 0.91, "grad_norm": 0.4711022973060608, "learning_rate": 1.9505851755526658e-05, "loss": 0.8654, "step": 774 }, { "epoch": 0.91, "grad_norm": 0.5381599068641663, "learning_rate": 1.9245773732119637e-05, "loss": 1.0086, "step": 775 }, { "epoch": 0.91, "grad_norm": 0.3786610960960388, "learning_rate": 1.8985695708712612e-05, "loss": 0.8949, "step": 776 }, { "epoch": 0.92, "grad_norm": 0.46414411067962646, "learning_rate": 1.8725617685305595e-05, "loss": 0.8106, "step": 777 }, { "epoch": 0.92, "grad_norm": 0.4330191910266876, "learning_rate": 1.846553966189857e-05, "loss": 1.0437, "step": 778 }, { "epoch": 0.92, "grad_norm": 0.6182777285575867, "learning_rate": 1.820546163849155e-05, "loss": 1.0076, "step": 779 }, { "epoch": 0.92, "grad_norm": 0.4420313239097595, "learning_rate": 1.7945383615084528e-05, "loss": 1.0418, "step": 780 }, { "epoch": 0.92, "grad_norm": 0.3915885090827942, "learning_rate": 1.7685305591677503e-05, "loss": 0.8567, "step": 781 }, { "epoch": 0.92, "grad_norm": 0.4120723605155945, "learning_rate": 1.7425227568270482e-05, "loss": 0.823, "step": 782 }, { "epoch": 0.92, "grad_norm": 0.6806041598320007, "learning_rate": 1.716514954486346e-05, "loss": 1.0303, "step": 783 }, { "epoch": 0.92, "grad_norm": 0.4460642635822296, "learning_rate": 1.6905071521456436e-05, "loss": 0.8871, "step": 784 }, { "epoch": 0.92, "grad_norm": 0.3997115194797516, "learning_rate": 1.6644993498049415e-05, "loss": 0.6781, "step": 785 }, { "epoch": 0.93, "grad_norm": 0.3535546660423279, "learning_rate": 1.6384915474642394e-05, "loss": 1.0946, "step": 786 }, { "epoch": 0.93, "grad_norm": 0.5428568720817566, "learning_rate": 1.6124837451235372e-05, "loss": 1.0807, "step": 787 }, { "epoch": 0.93, "grad_norm": 0.54714435338974, "learning_rate": 1.5864759427828348e-05, "loss": 0.7252, "step": 788 }, { "epoch": 0.93, "grad_norm": 0.4516313374042511, "learning_rate": 1.560468140442133e-05, "loss": 1.105, "step": 789 }, { "epoch": 0.93, "grad_norm": 0.35675543546676636, "learning_rate": 1.5344603381014305e-05, "loss": 0.891, "step": 790 }, { "epoch": 0.93, "grad_norm": 0.6124924421310425, "learning_rate": 1.5084525357607283e-05, "loss": 0.9851, "step": 791 }, { "epoch": 0.93, "grad_norm": 0.4543600082397461, "learning_rate": 1.4824447334200261e-05, "loss": 1.1088, "step": 792 }, { "epoch": 0.93, "grad_norm": 0.47420981526374817, "learning_rate": 1.4564369310793238e-05, "loss": 0.9064, "step": 793 }, { "epoch": 0.94, "grad_norm": 0.4191552400588989, "learning_rate": 1.4304291287386217e-05, "loss": 0.9856, "step": 794 }, { "epoch": 0.94, "grad_norm": 0.568281352519989, "learning_rate": 1.4044213263979194e-05, "loss": 0.7938, "step": 795 }, { "epoch": 0.94, "grad_norm": 0.38440269231796265, "learning_rate": 1.3784135240572171e-05, "loss": 0.8005, "step": 796 }, { "epoch": 0.94, "grad_norm": 0.4899348318576813, "learning_rate": 1.352405721716515e-05, "loss": 1.0936, "step": 797 }, { "epoch": 0.94, "grad_norm": 0.7555168271064758, "learning_rate": 1.3263979193758127e-05, "loss": 1.1329, "step": 798 }, { "epoch": 0.94, "grad_norm": 0.864571213722229, "learning_rate": 1.3003901170351108e-05, "loss": 1.2467, "step": 799 }, { "epoch": 0.94, "grad_norm": 0.3644244372844696, "learning_rate": 1.2743823146944083e-05, "loss": 0.8284, "step": 800 }, { "epoch": 0.94, "eval_loss": 0.9384466409683228, "eval_runtime": 547.1409, "eval_samples_per_second": 2.76, "eval_steps_per_second": 1.38, "step": 800 }, { "epoch": 0.94, "grad_norm": 0.45532289147377014, "learning_rate": 1.2483745123537062e-05, "loss": 0.9802, "step": 801 }, { "epoch": 0.94, "grad_norm": 0.37298470735549927, "learning_rate": 1.222366710013004e-05, "loss": 0.8374, "step": 802 }, { "epoch": 0.95, "grad_norm": 0.38847512006759644, "learning_rate": 1.1963589076723016e-05, "loss": 0.8607, "step": 803 }, { "epoch": 0.95, "grad_norm": 0.8099129796028137, "learning_rate": 1.1703511053315995e-05, "loss": 1.1332, "step": 804 }, { "epoch": 0.95, "grad_norm": 0.6113103032112122, "learning_rate": 1.1443433029908974e-05, "loss": 0.8674, "step": 805 }, { "epoch": 0.95, "grad_norm": 0.4203117787837982, "learning_rate": 1.1183355006501951e-05, "loss": 0.9592, "step": 806 }, { "epoch": 0.95, "grad_norm": 0.40829524397850037, "learning_rate": 1.092327698309493e-05, "loss": 0.7957, "step": 807 }, { "epoch": 0.95, "grad_norm": 0.6728403568267822, "learning_rate": 1.0663198959687907e-05, "loss": 0.8761, "step": 808 }, { "epoch": 0.95, "grad_norm": 0.39383500814437866, "learning_rate": 1.0403120936280884e-05, "loss": 1.0087, "step": 809 }, { "epoch": 0.95, "grad_norm": 0.5986286997795105, "learning_rate": 1.0143042912873863e-05, "loss": 1.0358, "step": 810 }, { "epoch": 0.96, "grad_norm": 0.47003424167633057, "learning_rate": 9.88296488946684e-06, "loss": 0.8709, "step": 811 }, { "epoch": 0.96, "grad_norm": 0.5308876633644104, "learning_rate": 9.622886866059819e-06, "loss": 0.7184, "step": 812 }, { "epoch": 0.96, "grad_norm": 0.36506423354148865, "learning_rate": 9.362808842652797e-06, "loss": 0.8865, "step": 813 }, { "epoch": 0.96, "grad_norm": 0.47633713483810425, "learning_rate": 9.102730819245774e-06, "loss": 0.8499, "step": 814 }, { "epoch": 0.96, "grad_norm": 0.5352996587753296, "learning_rate": 8.842652795838752e-06, "loss": 0.909, "step": 815 }, { "epoch": 0.96, "grad_norm": 0.5445082783699036, "learning_rate": 8.58257477243173e-06, "loss": 0.8987, "step": 816 }, { "epoch": 0.96, "grad_norm": 0.42454373836517334, "learning_rate": 8.322496749024707e-06, "loss": 0.8066, "step": 817 }, { "epoch": 0.96, "grad_norm": 0.39146989583969116, "learning_rate": 8.062418725617686e-06, "loss": 0.889, "step": 818 }, { "epoch": 0.96, "grad_norm": 0.6935787200927734, "learning_rate": 7.802340702210665e-06, "loss": 0.8192, "step": 819 }, { "epoch": 0.97, "grad_norm": 0.47039029002189636, "learning_rate": 7.542262678803641e-06, "loss": 1.0868, "step": 820 }, { "epoch": 0.97, "grad_norm": 0.4815388321876526, "learning_rate": 7.282184655396619e-06, "loss": 0.8655, "step": 821 }, { "epoch": 0.97, "grad_norm": 0.5231131315231323, "learning_rate": 7.022106631989597e-06, "loss": 0.893, "step": 822 }, { "epoch": 0.97, "grad_norm": 0.4355279803276062, "learning_rate": 6.762028608582575e-06, "loss": 0.7722, "step": 823 }, { "epoch": 0.97, "grad_norm": 0.3778000771999359, "learning_rate": 6.501950585175554e-06, "loss": 0.8279, "step": 824 }, { "epoch": 0.97, "grad_norm": 0.48129695653915405, "learning_rate": 6.241872561768531e-06, "loss": 0.732, "step": 825 }, { "epoch": 0.97, "grad_norm": 0.5184707641601562, "learning_rate": 5.981794538361508e-06, "loss": 0.8304, "step": 826 }, { "epoch": 0.97, "grad_norm": 0.4617622196674347, "learning_rate": 5.721716514954487e-06, "loss": 0.9741, "step": 827 }, { "epoch": 0.98, "grad_norm": 0.3710864782333374, "learning_rate": 5.461638491547465e-06, "loss": 0.933, "step": 828 }, { "epoch": 0.98, "grad_norm": 0.4909525215625763, "learning_rate": 5.201560468140442e-06, "loss": 0.7656, "step": 829 }, { "epoch": 0.98, "grad_norm": 0.48572537302970886, "learning_rate": 4.94148244473342e-06, "loss": 0.9586, "step": 830 }, { "epoch": 0.98, "grad_norm": 0.49216291308403015, "learning_rate": 4.681404421326399e-06, "loss": 0.7407, "step": 831 }, { "epoch": 0.98, "grad_norm": 0.8878616094589233, "learning_rate": 4.421326397919376e-06, "loss": 0.9886, "step": 832 }, { "epoch": 0.98, "grad_norm": 0.7310166954994202, "learning_rate": 4.161248374512354e-06, "loss": 0.8837, "step": 833 }, { "epoch": 0.98, "grad_norm": 0.8883714079856873, "learning_rate": 3.9011703511053325e-06, "loss": 1.1189, "step": 834 }, { "epoch": 0.98, "grad_norm": 0.4345027208328247, "learning_rate": 3.6410923276983096e-06, "loss": 0.9681, "step": 835 }, { "epoch": 0.98, "grad_norm": 0.42684683203697205, "learning_rate": 3.3810143042912876e-06, "loss": 1.0545, "step": 836 }, { "epoch": 0.99, "grad_norm": 0.4396170675754547, "learning_rate": 3.1209362808842655e-06, "loss": 0.9133, "step": 837 }, { "epoch": 0.99, "grad_norm": 0.3466357886791229, "learning_rate": 2.8608582574772434e-06, "loss": 0.9995, "step": 838 }, { "epoch": 0.99, "grad_norm": 0.35970377922058105, "learning_rate": 2.600780234070221e-06, "loss": 0.7297, "step": 839 }, { "epoch": 0.99, "grad_norm": 0.3779386579990387, "learning_rate": 2.3407022106631993e-06, "loss": 0.9407, "step": 840 }, { "epoch": 0.99, "grad_norm": 0.4766659140586853, "learning_rate": 2.080624187256177e-06, "loss": 0.9772, "step": 841 }, { "epoch": 0.99, "grad_norm": 0.35604748129844666, "learning_rate": 1.8205461638491548e-06, "loss": 0.7713, "step": 842 }, { "epoch": 0.99, "grad_norm": 0.42413535714149475, "learning_rate": 1.5604681404421327e-06, "loss": 0.9633, "step": 843 }, { "epoch": 0.99, "grad_norm": 0.43833646178245544, "learning_rate": 1.3003901170351105e-06, "loss": 0.9522, "step": 844 }, { "epoch": 1.0, "grad_norm": 0.48562091588974, "learning_rate": 1.0403120936280884e-06, "loss": 0.8831, "step": 845 }, { "epoch": 1.0, "grad_norm": 0.45852476358413696, "learning_rate": 7.802340702210664e-07, "loss": 0.7953, "step": 846 }, { "epoch": 1.0, "grad_norm": 0.5953251719474792, "learning_rate": 5.201560468140442e-07, "loss": 1.0921, "step": 847 }, { "epoch": 1.0, "grad_norm": 0.5428547859191895, "learning_rate": 2.600780234070221e-07, "loss": 0.965, "step": 848 }, { "epoch": 1.0, "grad_norm": 0.41269373893737793, "learning_rate": 0.0, "loss": 1.0192, "step": 849 }, { "epoch": 1.0, "step": 849, "total_flos": 3.9571832379644314e+17, "train_loss": 0.9633775743775429, "train_runtime": 18851.463, "train_samples_per_second": 0.72, "train_steps_per_second": 0.045 } ], "logging_steps": 1.0, "max_steps": 849, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "total_flos": 3.9571832379644314e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }