diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,4552 @@ +{ + "best_metric": 2.3818840980529785, + "best_model_checkpoint": "/ssd1/geonmin.kim/shortened-llm/outputs/phi2_1.8b_alpaca_enzh_fullparam/checkpoint-5800", + "epoch": 59.316075423457974, + "eval_steps": 100, + "global_step": 5800, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.010226909555768616, + "grad_norm": 3.242844581604004, + "learning_rate": 2.0000000000000002e-07, + "loss": 2.0435, + "step": 1 + }, + { + "epoch": 0.10226909555768617, + "grad_norm": 2.2696735858917236, + "learning_rate": 2.0000000000000003e-06, + "loss": 2.0397, + "step": 10 + }, + { + "epoch": 0.20453819111537233, + "grad_norm": 1.283516764640808, + "learning_rate": 4.000000000000001e-06, + "loss": 1.7851, + "step": 20 + }, + { + "epoch": 0.3068072866730585, + "grad_norm": 1.1238912343978882, + "learning_rate": 6e-06, + "loss": 1.6526, + "step": 30 + }, + { + "epoch": 0.40907638223074466, + "grad_norm": 1.0196110010147095, + "learning_rate": 8.000000000000001e-06, + "loss": 1.5391, + "step": 40 + }, + { + "epoch": 0.5113454777884308, + "grad_norm": 0.9775912761688232, + "learning_rate": 1e-05, + "loss": 1.4779, + "step": 50 + }, + { + "epoch": 0.613614573346117, + "grad_norm": 0.9060685038566589, + "learning_rate": 1.2e-05, + "loss": 1.4218, + "step": 60 + }, + { + "epoch": 0.7158836689038032, + "grad_norm": 0.9413829445838928, + "learning_rate": 1.4e-05, + "loss": 1.3978, + "step": 70 + }, + { + "epoch": 0.8181527644614893, + "grad_norm": 1.007612705230713, + "learning_rate": 1.6000000000000003e-05, + "loss": 1.3779, + "step": 80 + }, + { + "epoch": 0.9204218600191755, + "grad_norm": 0.9302048087120056, + "learning_rate": 1.8e-05, + "loss": 1.3387, + "step": 90 + }, + { + "epoch": 1.0226909555768615, + "grad_norm": 0.82750403881073, + "learning_rate": 2e-05, + "loss": 1.3056, + "step": 100 + }, + { + "epoch": 1.0226909555768615, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.201456069946289, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7448, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.41, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.051, + "step": 100 + }, + { + "epoch": 1.124960051134548, + "grad_norm": 0.8751218318939209, + "learning_rate": 1.9979166666666667e-05, + "loss": 1.2376, + "step": 110 + }, + { + "epoch": 1.2272291466922338, + "grad_norm": 0.8453890681266785, + "learning_rate": 1.9958333333333335e-05, + "loss": 1.2063, + "step": 120 + }, + { + "epoch": 1.3294982422499202, + "grad_norm": 0.8156936764717102, + "learning_rate": 1.99375e-05, + "loss": 1.2022, + "step": 130 + }, + { + "epoch": 1.4317673378076063, + "grad_norm": 0.8538714051246643, + "learning_rate": 1.991666666666667e-05, + "loss": 1.2052, + "step": 140 + }, + { + "epoch": 1.5340364333652925, + "grad_norm": 0.8687230944633484, + "learning_rate": 1.9895833333333334e-05, + "loss": 1.1913, + "step": 150 + }, + { + "epoch": 1.6363055289229784, + "grad_norm": 0.826032280921936, + "learning_rate": 1.9875000000000002e-05, + "loss": 1.1828, + "step": 160 + }, + { + "epoch": 1.7385746244806648, + "grad_norm": 0.8092572093009949, + "learning_rate": 1.9854166666666667e-05, + "loss": 1.1746, + "step": 170 + }, + { + "epoch": 1.840843720038351, + "grad_norm": 0.825042188167572, + "learning_rate": 1.9833333333333335e-05, + "loss": 1.169, + "step": 180 + }, + { + "epoch": 1.943112815596037, + "grad_norm": 0.8175085186958313, + "learning_rate": 1.98125e-05, + "loss": 1.166, + "step": 190 + }, + { + "epoch": 2.045381911153723, + "grad_norm": 0.8639950752258301, + "learning_rate": 1.979166666666667e-05, + "loss": 1.1158, + "step": 200 + }, + { + "epoch": 2.045381911153723, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1383644342422485, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7181, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.534, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.067, + "step": 200 + }, + { + "epoch": 2.1476510067114094, + "grad_norm": 0.7958649396896362, + "learning_rate": 1.9770833333333334e-05, + "loss": 1.0339, + "step": 210 + }, + { + "epoch": 2.249920102269096, + "grad_norm": 0.8010162711143494, + "learning_rate": 1.9750000000000002e-05, + "loss": 1.0426, + "step": 220 + }, + { + "epoch": 2.3521891978267817, + "grad_norm": 0.7885578274726868, + "learning_rate": 1.9729166666666667e-05, + "loss": 1.0432, + "step": 230 + }, + { + "epoch": 2.4544582933844676, + "grad_norm": 0.8269402980804443, + "learning_rate": 1.9708333333333336e-05, + "loss": 1.0379, + "step": 240 + }, + { + "epoch": 2.556727388942154, + "grad_norm": 0.8139219880104065, + "learning_rate": 1.96875e-05, + "loss": 1.0305, + "step": 250 + }, + { + "epoch": 2.6589964844998404, + "grad_norm": 0.8240389227867126, + "learning_rate": 1.9666666666666666e-05, + "loss": 1.032, + "step": 260 + }, + { + "epoch": 2.7612655800575263, + "grad_norm": 0.7902543544769287, + "learning_rate": 1.9645833333333334e-05, + "loss": 1.0203, + "step": 270 + }, + { + "epoch": 2.8635346756152127, + "grad_norm": 0.7832587361335754, + "learning_rate": 1.9625e-05, + "loss": 1.0266, + "step": 280 + }, + { + "epoch": 2.9658037711728986, + "grad_norm": 0.8064606189727783, + "learning_rate": 1.9604166666666668e-05, + "loss": 1.0292, + "step": 290 + }, + { + "epoch": 3.068072866730585, + "grad_norm": 0.8083460927009583, + "learning_rate": 1.9583333333333333e-05, + "loss": 0.9461, + "step": 300 + }, + { + "epoch": 3.068072866730585, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1164577007293701, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7799, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.247, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.031, + "step": 300 + }, + { + "epoch": 3.170341962288271, + "grad_norm": 0.7945894598960876, + "learning_rate": 1.95625e-05, + "loss": 0.9071, + "step": 310 + }, + { + "epoch": 3.2726110578459573, + "grad_norm": 0.8095683455467224, + "learning_rate": 1.9541666666666666e-05, + "loss": 0.9081, + "step": 320 + }, + { + "epoch": 3.3748801534036432, + "grad_norm": 0.779909074306488, + "learning_rate": 1.9520833333333335e-05, + "loss": 0.9171, + "step": 330 + }, + { + "epoch": 3.4771492489613296, + "grad_norm": 0.8173850774765015, + "learning_rate": 1.95e-05, + "loss": 0.9085, + "step": 340 + }, + { + "epoch": 3.5794183445190155, + "grad_norm": 0.8204581141471863, + "learning_rate": 1.9479166666666668e-05, + "loss": 0.9073, + "step": 350 + }, + { + "epoch": 3.681687440076702, + "grad_norm": 0.7985939383506775, + "learning_rate": 1.9458333333333333e-05, + "loss": 0.9181, + "step": 360 + }, + { + "epoch": 3.783956535634388, + "grad_norm": 0.8140767812728882, + "learning_rate": 1.94375e-05, + "loss": 0.9193, + "step": 370 + }, + { + "epoch": 3.886225631192074, + "grad_norm": 0.8102839589118958, + "learning_rate": 1.9416666666666667e-05, + "loss": 0.9157, + "step": 380 + }, + { + "epoch": 3.98849472674976, + "grad_norm": 0.7886559963226318, + "learning_rate": 1.9395833333333335e-05, + "loss": 0.9191, + "step": 390 + }, + { + "epoch": 4.090763822307446, + "grad_norm": 0.7990887761116028, + "learning_rate": 1.9375e-05, + "loss": 0.8203, + "step": 400 + }, + { + "epoch": 4.090763822307446, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1279124021530151, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7222, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.515, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.064, + "step": 400 + }, + { + "epoch": 4.193032917865133, + "grad_norm": 0.8134937286376953, + "learning_rate": 1.935416666666667e-05, + "loss": 0.81, + "step": 410 + }, + { + "epoch": 4.295302013422819, + "grad_norm": 0.8099086284637451, + "learning_rate": 1.9333333333333333e-05, + "loss": 0.8, + "step": 420 + }, + { + "epoch": 4.397571108980505, + "grad_norm": 0.7907654047012329, + "learning_rate": 1.9312500000000002e-05, + "loss": 0.8025, + "step": 430 + }, + { + "epoch": 4.499840204538192, + "grad_norm": 0.835167407989502, + "learning_rate": 1.9291666666666667e-05, + "loss": 0.804, + "step": 440 + }, + { + "epoch": 4.6021093000958775, + "grad_norm": 0.8259727358818054, + "learning_rate": 1.9270833333333335e-05, + "loss": 0.8069, + "step": 450 + }, + { + "epoch": 4.704378395653563, + "grad_norm": 0.8095329999923706, + "learning_rate": 1.925e-05, + "loss": 0.8099, + "step": 460 + }, + { + "epoch": 4.806647491211249, + "grad_norm": 0.8239871859550476, + "learning_rate": 1.922916666666667e-05, + "loss": 0.8178, + "step": 470 + }, + { + "epoch": 4.908916586768935, + "grad_norm": 0.8325999975204468, + "learning_rate": 1.9208333333333337e-05, + "loss": 0.8077, + "step": 480 + }, + { + "epoch": 5.011185682326622, + "grad_norm": 0.813887357711792, + "learning_rate": 1.9187500000000002e-05, + "loss": 0.8061, + "step": 490 + }, + { + "epoch": 5.113454777884308, + "grad_norm": 0.8288919925689697, + "learning_rate": 1.916666666666667e-05, + "loss": 0.7063, + "step": 500 + }, + { + "epoch": 5.113454777884308, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1451531648635864, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7807, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.243, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.03, + "step": 500 + }, + { + "epoch": 5.215723873441994, + "grad_norm": 0.8298630118370056, + "learning_rate": 1.9145833333333336e-05, + "loss": 0.7049, + "step": 510 + }, + { + "epoch": 5.317992968999681, + "grad_norm": 0.8246304988861084, + "learning_rate": 1.9125000000000004e-05, + "loss": 0.7104, + "step": 520 + }, + { + "epoch": 5.420262064557367, + "grad_norm": 0.7879628539085388, + "learning_rate": 1.910416666666667e-05, + "loss": 0.7147, + "step": 530 + }, + { + "epoch": 5.522531160115053, + "grad_norm": 0.8137240409851074, + "learning_rate": 1.9083333333333338e-05, + "loss": 0.7079, + "step": 540 + }, + { + "epoch": 5.624800255672739, + "grad_norm": 0.8201763033866882, + "learning_rate": 1.9062500000000003e-05, + "loss": 0.7162, + "step": 550 + }, + { + "epoch": 5.727069351230425, + "grad_norm": 0.8055077791213989, + "learning_rate": 1.9041666666666668e-05, + "loss": 0.7241, + "step": 560 + }, + { + "epoch": 5.829338446788111, + "grad_norm": 0.8504828810691833, + "learning_rate": 1.9020833333333336e-05, + "loss": 0.7241, + "step": 570 + }, + { + "epoch": 5.931607542345797, + "grad_norm": 0.8135042190551758, + "learning_rate": 1.9e-05, + "loss": 0.7276, + "step": 580 + }, + { + "epoch": 6.033876637903483, + "grad_norm": 0.907433271408081, + "learning_rate": 1.897916666666667e-05, + "loss": 0.6957, + "step": 590 + }, + { + "epoch": 6.13614573346117, + "grad_norm": 0.8348938822746277, + "learning_rate": 1.8958333333333334e-05, + "loss": 0.6172, + "step": 600 + }, + { + "epoch": 6.13614573346117, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.171920657157898, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7497, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.387, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.048, + "step": 600 + }, + { + "epoch": 6.238414829018856, + "grad_norm": 0.8306870460510254, + "learning_rate": 1.8937500000000003e-05, + "loss": 0.6255, + "step": 610 + }, + { + "epoch": 6.340683924576542, + "grad_norm": 0.8185003399848938, + "learning_rate": 1.8916666666666668e-05, + "loss": 0.6218, + "step": 620 + }, + { + "epoch": 6.442953020134228, + "grad_norm": 0.8396403789520264, + "learning_rate": 1.8895833333333336e-05, + "loss": 0.6282, + "step": 630 + }, + { + "epoch": 6.545222115691915, + "grad_norm": 0.8167343139648438, + "learning_rate": 1.8875e-05, + "loss": 0.6305, + "step": 640 + }, + { + "epoch": 6.6474912112496005, + "grad_norm": 0.8276931047439575, + "learning_rate": 1.885416666666667e-05, + "loss": 0.6378, + "step": 650 + }, + { + "epoch": 6.7497603068072864, + "grad_norm": 0.8405306339263916, + "learning_rate": 1.8833333333333335e-05, + "loss": 0.6375, + "step": 660 + }, + { + "epoch": 6.852029402364973, + "grad_norm": 0.8206018805503845, + "learning_rate": 1.8812500000000003e-05, + "loss": 0.6393, + "step": 670 + }, + { + "epoch": 6.954298497922659, + "grad_norm": 0.8008025884628296, + "learning_rate": 1.8791666666666668e-05, + "loss": 0.6448, + "step": 680 + }, + { + "epoch": 7.056567593480345, + "grad_norm": 0.8231362104415894, + "learning_rate": 1.8770833333333337e-05, + "loss": 0.5929, + "step": 690 + }, + { + "epoch": 7.158836689038031, + "grad_norm": 0.8469756841659546, + "learning_rate": 1.8750000000000002e-05, + "loss": 0.5392, + "step": 700 + }, + { + "epoch": 7.158836689038031, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.2256364822387695, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8506, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.92, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.99, + "step": 700 + }, + { + "epoch": 7.261105784595718, + "grad_norm": 0.8122310042381287, + "learning_rate": 1.8729166666666667e-05, + "loss": 0.54, + "step": 710 + }, + { + "epoch": 7.363374880153404, + "grad_norm": 0.8621034026145935, + "learning_rate": 1.8708333333333335e-05, + "loss": 0.5534, + "step": 720 + }, + { + "epoch": 7.46564397571109, + "grad_norm": 0.8529708981513977, + "learning_rate": 1.86875e-05, + "loss": 0.5516, + "step": 730 + }, + { + "epoch": 7.567913071268776, + "grad_norm": 0.8307532072067261, + "learning_rate": 1.866666666666667e-05, + "loss": 0.5531, + "step": 740 + }, + { + "epoch": 7.6701821668264625, + "grad_norm": 0.8403590321540833, + "learning_rate": 1.8645833333333334e-05, + "loss": 0.5644, + "step": 750 + }, + { + "epoch": 7.772451262384148, + "grad_norm": 0.8332897424697876, + "learning_rate": 1.8625000000000002e-05, + "loss": 0.5607, + "step": 760 + }, + { + "epoch": 7.874720357941834, + "grad_norm": 0.866201639175415, + "learning_rate": 1.8604166666666667e-05, + "loss": 0.5658, + "step": 770 + }, + { + "epoch": 7.97698945349952, + "grad_norm": 0.8362734913825989, + "learning_rate": 1.8583333333333336e-05, + "loss": 0.566, + "step": 780 + }, + { + "epoch": 8.079258549057206, + "grad_norm": 0.8327781558036804, + "learning_rate": 1.85625e-05, + "loss": 0.4963, + "step": 790 + }, + { + "epoch": 8.181527644614892, + "grad_norm": 0.827268123626709, + "learning_rate": 1.854166666666667e-05, + "loss": 0.4752, + "step": 800 + }, + { + "epoch": 8.181527644614892, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.2643427848815918, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.893, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.726, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.966, + "step": 800 + }, + { + "epoch": 8.28379674017258, + "grad_norm": 0.7902690172195435, + "learning_rate": 1.8520833333333334e-05, + "loss": 0.4843, + "step": 810 + }, + { + "epoch": 8.386065835730266, + "grad_norm": 0.8375966548919678, + "learning_rate": 1.8500000000000002e-05, + "loss": 0.4783, + "step": 820 + }, + { + "epoch": 8.488334931287952, + "grad_norm": 0.8248066306114197, + "learning_rate": 1.8479166666666667e-05, + "loss": 0.4785, + "step": 830 + }, + { + "epoch": 8.590604026845638, + "grad_norm": 0.8647124171257019, + "learning_rate": 1.8458333333333336e-05, + "loss": 0.4874, + "step": 840 + }, + { + "epoch": 8.692873122403324, + "grad_norm": 0.8471198678016663, + "learning_rate": 1.84375e-05, + "loss": 0.4884, + "step": 850 + }, + { + "epoch": 8.79514221796101, + "grad_norm": 0.8394653797149658, + "learning_rate": 1.8416666666666666e-05, + "loss": 0.497, + "step": 860 + }, + { + "epoch": 8.897411313518695, + "grad_norm": 0.8621006011962891, + "learning_rate": 1.8395833333333334e-05, + "loss": 0.4923, + "step": 870 + }, + { + "epoch": 8.999680409076383, + "grad_norm": 0.830193817615509, + "learning_rate": 1.8375e-05, + "loss": 0.4917, + "step": 880 + }, + { + "epoch": 9.101949504634069, + "grad_norm": 0.8418065309524536, + "learning_rate": 1.8354166666666668e-05, + "loss": 0.4104, + "step": 890 + }, + { + "epoch": 9.204218600191755, + "grad_norm": 0.8672240376472473, + "learning_rate": 1.8333333333333333e-05, + "loss": 0.4165, + "step": 900 + }, + { + "epoch": 9.204218600191755, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.2995867729187012, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8906, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.737, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.967, + "step": 900 + }, + { + "epoch": 9.306487695749441, + "grad_norm": 0.8118753433227539, + "learning_rate": 1.83125e-05, + "loss": 0.4181, + "step": 910 + }, + { + "epoch": 9.408756791307127, + "grad_norm": 0.8521497845649719, + "learning_rate": 1.8291666666666666e-05, + "loss": 0.4202, + "step": 920 + }, + { + "epoch": 9.511025886864813, + "grad_norm": 0.8396993279457092, + "learning_rate": 1.8270833333333335e-05, + "loss": 0.4247, + "step": 930 + }, + { + "epoch": 9.613294982422499, + "grad_norm": 0.8380371332168579, + "learning_rate": 1.825e-05, + "loss": 0.427, + "step": 940 + }, + { + "epoch": 9.715564077980185, + "grad_norm": 0.819240927696228, + "learning_rate": 1.8229166666666668e-05, + "loss": 0.4271, + "step": 950 + }, + { + "epoch": 9.817833173537872, + "grad_norm": 0.8467490077018738, + "learning_rate": 1.8208333333333333e-05, + "loss": 0.4288, + "step": 960 + }, + { + "epoch": 9.920102269095558, + "grad_norm": 0.8679558634757996, + "learning_rate": 1.81875e-05, + "loss": 0.4283, + "step": 970 + }, + { + "epoch": 10.022371364653244, + "grad_norm": 0.8336887359619141, + "learning_rate": 1.8166666666666667e-05, + "loss": 0.4244, + "step": 980 + }, + { + "epoch": 10.12464046021093, + "grad_norm": 0.8790706396102905, + "learning_rate": 1.8145833333333335e-05, + "loss": 0.3627, + "step": 990 + }, + { + "epoch": 10.226909555768616, + "grad_norm": 0.8260719180107117, + "learning_rate": 1.8125e-05, + "loss": 0.3622, + "step": 1000 + }, + { + "epoch": 10.226909555768616, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.3518736362457275, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.891, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.735, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.967, + "step": 1000 + }, + { + "epoch": 10.329178651326302, + "grad_norm": 0.8354102969169617, + "learning_rate": 1.810416666666667e-05, + "loss": 0.3654, + "step": 1010 + }, + { + "epoch": 10.431447746883988, + "grad_norm": 0.833742618560791, + "learning_rate": 1.8083333333333334e-05, + "loss": 0.3675, + "step": 1020 + }, + { + "epoch": 10.533716842441674, + "grad_norm": 0.814929723739624, + "learning_rate": 1.8062500000000002e-05, + "loss": 0.368, + "step": 1030 + }, + { + "epoch": 10.635985937999362, + "grad_norm": 0.8500260710716248, + "learning_rate": 1.8041666666666667e-05, + "loss": 0.3704, + "step": 1040 + }, + { + "epoch": 10.738255033557047, + "grad_norm": 0.8280666470527649, + "learning_rate": 1.8020833333333335e-05, + "loss": 0.3722, + "step": 1050 + }, + { + "epoch": 10.840524129114733, + "grad_norm": 0.871941089630127, + "learning_rate": 1.8e-05, + "loss": 0.3768, + "step": 1060 + }, + { + "epoch": 10.94279322467242, + "grad_norm": 0.8509662747383118, + "learning_rate": 1.797916666666667e-05, + "loss": 0.3829, + "step": 1070 + }, + { + "epoch": 11.045062320230105, + "grad_norm": 0.8621676564216614, + "learning_rate": 1.7958333333333334e-05, + "loss": 0.3512, + "step": 1080 + }, + { + "epoch": 11.147331415787791, + "grad_norm": 0.8415457010269165, + "learning_rate": 1.7937500000000002e-05, + "loss": 0.3092, + "step": 1090 + }, + { + "epoch": 11.249600511345477, + "grad_norm": 0.8087013363838196, + "learning_rate": 1.7916666666666667e-05, + "loss": 0.3158, + "step": 1100 + }, + { + "epoch": 11.249600511345477, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.4024713039398193, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8756, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.805, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.976, + "step": 1100 + }, + { + "epoch": 11.351869606903165, + "grad_norm": 0.8239210844039917, + "learning_rate": 1.7895833333333336e-05, + "loss": 0.3189, + "step": 1110 + }, + { + "epoch": 11.45413870246085, + "grad_norm": 0.8607499003410339, + "learning_rate": 1.7875e-05, + "loss": 0.3222, + "step": 1120 + }, + { + "epoch": 11.556407798018537, + "grad_norm": 0.8538540601730347, + "learning_rate": 1.785416666666667e-05, + "loss": 0.3207, + "step": 1130 + }, + { + "epoch": 11.658676893576223, + "grad_norm": 0.8388919234275818, + "learning_rate": 1.7833333333333334e-05, + "loss": 0.3308, + "step": 1140 + }, + { + "epoch": 11.760945989133909, + "grad_norm": 0.8096144795417786, + "learning_rate": 1.7812500000000003e-05, + "loss": 0.3256, + "step": 1150 + }, + { + "epoch": 11.863215084691594, + "grad_norm": 0.8199524879455566, + "learning_rate": 1.7791666666666668e-05, + "loss": 0.3272, + "step": 1160 + }, + { + "epoch": 11.96548418024928, + "grad_norm": 0.8503059148788452, + "learning_rate": 1.7770833333333336e-05, + "loss": 0.3313, + "step": 1170 + }, + { + "epoch": 12.067753275806966, + "grad_norm": 0.8376160860061646, + "learning_rate": 1.775e-05, + "loss": 0.2907, + "step": 1180 + }, + { + "epoch": 12.170022371364654, + "grad_norm": 5.976499557495117, + "learning_rate": 1.772916666666667e-05, + "loss": 0.2719, + "step": 1190 + }, + { + "epoch": 12.27229146692234, + "grad_norm": 0.8097366094589233, + "learning_rate": 1.7708333333333335e-05, + "loss": 0.2782, + "step": 1200 + }, + { + "epoch": 12.27229146692234, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.4481781721115112, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.9199, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.603, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.95, + "step": 1200 + }, + { + "epoch": 12.374560562480026, + "grad_norm": 0.7708520293235779, + "learning_rate": 1.7687500000000003e-05, + "loss": 0.2749, + "step": 1210 + }, + { + "epoch": 12.476829658037712, + "grad_norm": 0.8598915338516235, + "learning_rate": 1.7666666666666668e-05, + "loss": 0.2784, + "step": 1220 + }, + { + "epoch": 12.579098753595398, + "grad_norm": 0.8157161474227905, + "learning_rate": 1.7645833333333336e-05, + "loss": 0.2794, + "step": 1230 + }, + { + "epoch": 12.681367849153084, + "grad_norm": 0.8496010899543762, + "learning_rate": 1.7625e-05, + "loss": 0.2825, + "step": 1240 + }, + { + "epoch": 12.78363694471077, + "grad_norm": 0.815390944480896, + "learning_rate": 1.760416666666667e-05, + "loss": 0.2848, + "step": 1250 + }, + { + "epoch": 12.885906040268456, + "grad_norm": 0.8204165697097778, + "learning_rate": 1.7583333333333335e-05, + "loss": 0.2883, + "step": 1260 + }, + { + "epoch": 12.988175135826143, + "grad_norm": 0.8374896049499512, + "learning_rate": 1.7562500000000003e-05, + "loss": 0.2909, + "step": 1270 + }, + { + "epoch": 13.09044423138383, + "grad_norm": 0.7642439007759094, + "learning_rate": 1.754166666666667e-05, + "loss": 0.2406, + "step": 1280 + }, + { + "epoch": 13.192713326941515, + "grad_norm": 0.7860681414604187, + "learning_rate": 1.7520833333333337e-05, + "loss": 0.2358, + "step": 1290 + }, + { + "epoch": 13.294982422499201, + "grad_norm": 0.8119321465492249, + "learning_rate": 1.7500000000000002e-05, + "loss": 0.2357, + "step": 1300 + }, + { + "epoch": 13.294982422499201, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.4959287643432617, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8771, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.799, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.975, + "step": 1300 + }, + { + "epoch": 13.397251518056887, + "grad_norm": 0.7898840308189392, + "learning_rate": 1.7479166666666667e-05, + "loss": 0.2409, + "step": 1310 + }, + { + "epoch": 13.499520613614573, + "grad_norm": 0.8298600912094116, + "learning_rate": 1.7458333333333335e-05, + "loss": 0.2399, + "step": 1320 + }, + { + "epoch": 13.601789709172259, + "grad_norm": 0.7994723916053772, + "learning_rate": 1.74375e-05, + "loss": 0.2468, + "step": 1330 + }, + { + "epoch": 13.704058804729947, + "grad_norm": 0.822475790977478, + "learning_rate": 1.741666666666667e-05, + "loss": 0.2481, + "step": 1340 + }, + { + "epoch": 13.806327900287632, + "grad_norm": 0.8012453317642212, + "learning_rate": 1.7395833333333334e-05, + "loss": 0.2473, + "step": 1350 + }, + { + "epoch": 13.908596995845318, + "grad_norm": 0.8046063780784607, + "learning_rate": 1.7375000000000002e-05, + "loss": 0.2532, + "step": 1360 + }, + { + "epoch": 14.010866091403004, + "grad_norm": 0.7180681228637695, + "learning_rate": 1.7354166666666667e-05, + "loss": 0.2451, + "step": 1370 + }, + { + "epoch": 14.11313518696069, + "grad_norm": 0.7648767828941345, + "learning_rate": 1.7333333333333336e-05, + "loss": 0.2026, + "step": 1380 + }, + { + "epoch": 14.215404282518376, + "grad_norm": 0.782311201095581, + "learning_rate": 1.73125e-05, + "loss": 0.2062, + "step": 1390 + }, + { + "epoch": 14.317673378076062, + "grad_norm": 0.7766838669776917, + "learning_rate": 1.729166666666667e-05, + "loss": 0.2079, + "step": 1400 + }, + { + "epoch": 14.317673378076062, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.535245418548584, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8983, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.701, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.963, + "step": 1400 + }, + { + "epoch": 14.419942473633748, + "grad_norm": 0.7595117688179016, + "learning_rate": 1.7270833333333334e-05, + "loss": 0.2097, + "step": 1410 + }, + { + "epoch": 14.522211569191436, + "grad_norm": 0.7640486359596252, + "learning_rate": 1.7250000000000003e-05, + "loss": 0.2107, + "step": 1420 + }, + { + "epoch": 14.624480664749122, + "grad_norm": 0.8100217580795288, + "learning_rate": 1.7229166666666668e-05, + "loss": 0.2132, + "step": 1430 + }, + { + "epoch": 14.726749760306808, + "grad_norm": 0.7824357748031616, + "learning_rate": 1.7208333333333336e-05, + "loss": 0.2124, + "step": 1440 + }, + { + "epoch": 14.829018855864494, + "grad_norm": 0.8215783834457397, + "learning_rate": 1.71875e-05, + "loss": 0.2172, + "step": 1450 + }, + { + "epoch": 14.93128795142218, + "grad_norm": 0.791244626045227, + "learning_rate": 1.7166666666666666e-05, + "loss": 0.2183, + "step": 1460 + }, + { + "epoch": 15.033557046979865, + "grad_norm": 0.7542420625686646, + "learning_rate": 1.7145833333333334e-05, + "loss": 0.2083, + "step": 1470 + }, + { + "epoch": 15.135826142537551, + "grad_norm": 0.7965226173400879, + "learning_rate": 1.7125e-05, + "loss": 0.1773, + "step": 1480 + }, + { + "epoch": 15.238095238095237, + "grad_norm": 0.764574408531189, + "learning_rate": 1.7104166666666668e-05, + "loss": 0.1775, + "step": 1490 + }, + { + "epoch": 15.340364333652925, + "grad_norm": 0.768020749092102, + "learning_rate": 1.7083333333333333e-05, + "loss": 0.181, + "step": 1500 + }, + { + "epoch": 15.340364333652925, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.57305908203125, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7396, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.434, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.054, + "step": 1500 + }, + { + "epoch": 15.44263342921061, + "grad_norm": 0.7567213773727417, + "learning_rate": 1.70625e-05, + "loss": 0.181, + "step": 1510 + }, + { + "epoch": 15.544902524768297, + "grad_norm": 0.7426446080207825, + "learning_rate": 1.7041666666666666e-05, + "loss": 0.1824, + "step": 1520 + }, + { + "epoch": 15.647171620325983, + "grad_norm": 0.750170111656189, + "learning_rate": 1.7020833333333335e-05, + "loss": 0.1855, + "step": 1530 + }, + { + "epoch": 15.749440715883669, + "grad_norm": 0.7680428624153137, + "learning_rate": 1.7e-05, + "loss": 0.1883, + "step": 1540 + }, + { + "epoch": 15.851709811441355, + "grad_norm": 0.8081603646278381, + "learning_rate": 1.6979166666666668e-05, + "loss": 0.1892, + "step": 1550 + }, + { + "epoch": 15.95397890699904, + "grad_norm": 0.7882938385009766, + "learning_rate": 1.6958333333333333e-05, + "loss": 0.1912, + "step": 1560 + }, + { + "epoch": 16.05624800255673, + "grad_norm": 0.7245915532112122, + "learning_rate": 1.6937500000000002e-05, + "loss": 0.1717, + "step": 1570 + }, + { + "epoch": 16.158517098114412, + "grad_norm": 0.7572883367538452, + "learning_rate": 1.6916666666666667e-05, + "loss": 0.1577, + "step": 1580 + }, + { + "epoch": 16.2607861936721, + "grad_norm": 0.7576362490653992, + "learning_rate": 1.6895833333333335e-05, + "loss": 0.157, + "step": 1590 + }, + { + "epoch": 16.363055289229784, + "grad_norm": 0.7182960510253906, + "learning_rate": 1.6875e-05, + "loss": 0.1555, + "step": 1600 + }, + { + "epoch": 16.363055289229784, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.6099534034729004, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8255, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.036, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.005, + "step": 1600 + }, + { + "epoch": 16.465324384787472, + "grad_norm": 0.7401660084724426, + "learning_rate": 1.685416666666667e-05, + "loss": 0.1573, + "step": 1610 + }, + { + "epoch": 16.56759348034516, + "grad_norm": 0.7596891522407532, + "learning_rate": 1.6833333333333334e-05, + "loss": 0.1589, + "step": 1620 + }, + { + "epoch": 16.669862575902844, + "grad_norm": 0.767320990562439, + "learning_rate": 1.6812500000000002e-05, + "loss": 0.1613, + "step": 1630 + }, + { + "epoch": 16.77213167146053, + "grad_norm": 0.7579568028450012, + "learning_rate": 1.6791666666666667e-05, + "loss": 0.1624, + "step": 1640 + }, + { + "epoch": 16.874400767018216, + "grad_norm": 0.7488529682159424, + "learning_rate": 1.6770833333333336e-05, + "loss": 0.1646, + "step": 1650 + }, + { + "epoch": 16.976669862575903, + "grad_norm": 0.7528676390647888, + "learning_rate": 1.675e-05, + "loss": 0.1686, + "step": 1660 + }, + { + "epoch": 17.078938958133588, + "grad_norm": 0.6930709481239319, + "learning_rate": 1.672916666666667e-05, + "loss": 0.1429, + "step": 1670 + }, + { + "epoch": 17.181208053691275, + "grad_norm": 0.7119250297546387, + "learning_rate": 1.6708333333333334e-05, + "loss": 0.1355, + "step": 1680 + }, + { + "epoch": 17.283477149248963, + "grad_norm": 0.6774548292160034, + "learning_rate": 1.6687500000000002e-05, + "loss": 0.1348, + "step": 1690 + }, + { + "epoch": 17.385746244806647, + "grad_norm": 0.7192063927650452, + "learning_rate": 1.6666666666666667e-05, + "loss": 0.1368, + "step": 1700 + }, + { + "epoch": 17.385746244806647, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.6659198999404907, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8832, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.771, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.971, + "step": 1700 + }, + { + "epoch": 17.488015340364335, + "grad_norm": 0.6981900930404663, + "learning_rate": 1.6645833333333336e-05, + "loss": 0.1379, + "step": 1710 + }, + { + "epoch": 17.59028443592202, + "grad_norm": 0.7018482685089111, + "learning_rate": 1.6625e-05, + "loss": 0.1413, + "step": 1720 + }, + { + "epoch": 17.692553531479707, + "grad_norm": 0.7282826900482178, + "learning_rate": 1.660416666666667e-05, + "loss": 0.1437, + "step": 1730 + }, + { + "epoch": 17.79482262703739, + "grad_norm": 0.751104474067688, + "learning_rate": 1.6583333333333334e-05, + "loss": 0.142, + "step": 1740 + }, + { + "epoch": 17.89709172259508, + "grad_norm": 0.7333133816719055, + "learning_rate": 1.6562500000000003e-05, + "loss": 0.1452, + "step": 1750 + }, + { + "epoch": 17.999360818152766, + "grad_norm": 0.7490417957305908, + "learning_rate": 1.6541666666666668e-05, + "loss": 0.1445, + "step": 1760 + }, + { + "epoch": 18.10162991371045, + "grad_norm": 0.7038053274154663, + "learning_rate": 1.6520833333333336e-05, + "loss": 0.1196, + "step": 1770 + }, + { + "epoch": 18.203899009268138, + "grad_norm": 0.6416111588478088, + "learning_rate": 1.65e-05, + "loss": 0.1189, + "step": 1780 + }, + { + "epoch": 18.306168104825822, + "grad_norm": 0.6799057126045227, + "learning_rate": 1.647916666666667e-05, + "loss": 0.1187, + "step": 1790 + }, + { + "epoch": 18.40843720038351, + "grad_norm": 0.70688396692276, + "learning_rate": 1.6458333333333335e-05, + "loss": 0.1195, + "step": 1800 + }, + { + "epoch": 18.40843720038351, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.701189398765564, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8449, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.947, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.993, + "step": 1800 + }, + { + "epoch": 18.510706295941194, + "grad_norm": 0.6970006823539734, + "learning_rate": 1.6437500000000003e-05, + "loss": 0.1223, + "step": 1810 + }, + { + "epoch": 18.612975391498882, + "grad_norm": 0.7127917408943176, + "learning_rate": 1.6416666666666668e-05, + "loss": 0.1228, + "step": 1820 + }, + { + "epoch": 18.715244487056566, + "grad_norm": 0.6875537633895874, + "learning_rate": 1.6395833333333337e-05, + "loss": 0.1234, + "step": 1830 + }, + { + "epoch": 18.817513582614254, + "grad_norm": 0.7249884009361267, + "learning_rate": 1.6375e-05, + "loss": 0.126, + "step": 1840 + }, + { + "epoch": 18.91978267817194, + "grad_norm": 0.7007323503494263, + "learning_rate": 1.635416666666667e-05, + "loss": 0.1243, + "step": 1850 + }, + { + "epoch": 19.022051773729626, + "grad_norm": 0.6626154184341431, + "learning_rate": 1.6333333333333335e-05, + "loss": 0.1233, + "step": 1860 + }, + { + "epoch": 19.124320869287313, + "grad_norm": 0.6599249243736267, + "learning_rate": 1.6312500000000003e-05, + "loss": 0.1034, + "step": 1870 + }, + { + "epoch": 19.226589964844997, + "grad_norm": 0.6787338256835938, + "learning_rate": 1.629166666666667e-05, + "loss": 0.1031, + "step": 1880 + }, + { + "epoch": 19.328859060402685, + "grad_norm": 0.6463894248008728, + "learning_rate": 1.6270833333333337e-05, + "loss": 0.106, + "step": 1890 + }, + { + "epoch": 19.43112815596037, + "grad_norm": 0.6649991869926453, + "learning_rate": 1.6250000000000002e-05, + "loss": 0.1063, + "step": 1900 + }, + { + "epoch": 19.43112815596037, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.7399890422821045, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8883, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.747, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.968, + "step": 1900 + }, + { + "epoch": 19.533397251518057, + "grad_norm": 0.6542336940765381, + "learning_rate": 1.6229166666666667e-05, + "loss": 0.1059, + "step": 1910 + }, + { + "epoch": 19.635666347075745, + "grad_norm": 0.6741412281990051, + "learning_rate": 1.6208333333333335e-05, + "loss": 0.1088, + "step": 1920 + }, + { + "epoch": 19.73793544263343, + "grad_norm": 0.6898075938224792, + "learning_rate": 1.61875e-05, + "loss": 0.1087, + "step": 1930 + }, + { + "epoch": 19.840204538191117, + "grad_norm": 0.6611754298210144, + "learning_rate": 1.616666666666667e-05, + "loss": 0.1098, + "step": 1940 + }, + { + "epoch": 19.9424736337488, + "grad_norm": 0.6986993551254272, + "learning_rate": 1.6145833333333334e-05, + "loss": 0.1114, + "step": 1950 + }, + { + "epoch": 20.04474272930649, + "grad_norm": 0.6218557953834534, + "learning_rate": 1.6125000000000002e-05, + "loss": 0.1028, + "step": 1960 + }, + { + "epoch": 20.147011824864173, + "grad_norm": 0.6389386653900146, + "learning_rate": 1.6104166666666667e-05, + "loss": 0.0922, + "step": 1970 + }, + { + "epoch": 20.24928092042186, + "grad_norm": 0.6259350776672363, + "learning_rate": 1.6083333333333336e-05, + "loss": 0.0924, + "step": 1980 + }, + { + "epoch": 20.351550015979548, + "grad_norm": 0.6429017782211304, + "learning_rate": 1.60625e-05, + "loss": 0.0935, + "step": 1990 + }, + { + "epoch": 20.453819111537232, + "grad_norm": 0.6409590244293213, + "learning_rate": 1.604166666666667e-05, + "loss": 0.0925, + "step": 2000 + }, + { + "epoch": 20.453819111537232, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.770869493484497, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8821, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.776, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.972, + "step": 2000 + }, + { + "epoch": 20.55608820709492, + "grad_norm": 0.6365486979484558, + "learning_rate": 1.6020833333333334e-05, + "loss": 0.0945, + "step": 2010 + }, + { + "epoch": 20.658357302652604, + "grad_norm": 0.6637431979179382, + "learning_rate": 1.6000000000000003e-05, + "loss": 0.0963, + "step": 2020 + }, + { + "epoch": 20.76062639821029, + "grad_norm": 0.6256012916564941, + "learning_rate": 1.5979166666666668e-05, + "loss": 0.0981, + "step": 2030 + }, + { + "epoch": 20.862895493767976, + "grad_norm": 0.6379542946815491, + "learning_rate": 1.5958333333333336e-05, + "loss": 0.0976, + "step": 2040 + }, + { + "epoch": 20.965164589325664, + "grad_norm": 0.6680212020874023, + "learning_rate": 1.59375e-05, + "loss": 0.0982, + "step": 2050 + }, + { + "epoch": 21.067433684883348, + "grad_norm": 0.6050080060958862, + "learning_rate": 1.5916666666666666e-05, + "loss": 0.0876, + "step": 2060 + }, + { + "epoch": 21.169702780441035, + "grad_norm": 0.6315256953239441, + "learning_rate": 1.5895833333333335e-05, + "loss": 0.0824, + "step": 2070 + }, + { + "epoch": 21.271971875998723, + "grad_norm": 0.6160369515419006, + "learning_rate": 1.5875e-05, + "loss": 0.0817, + "step": 2080 + }, + { + "epoch": 21.374240971556407, + "grad_norm": 0.5967450737953186, + "learning_rate": 1.5854166666666668e-05, + "loss": 0.0828, + "step": 2090 + }, + { + "epoch": 21.476510067114095, + "grad_norm": 0.6033092737197876, + "learning_rate": 1.5833333333333333e-05, + "loss": 0.0832, + "step": 2100 + }, + { + "epoch": 21.476510067114095, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.8033993244171143, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8632, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.862, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.983, + "step": 2100 + }, + { + "epoch": 21.57877916267178, + "grad_norm": 0.6197848916053772, + "learning_rate": 1.58125e-05, + "loss": 0.0852, + "step": 2110 + }, + { + "epoch": 21.681048258229467, + "grad_norm": 0.6332718133926392, + "learning_rate": 1.5791666666666667e-05, + "loss": 0.0853, + "step": 2120 + }, + { + "epoch": 21.78331735378715, + "grad_norm": 0.6289674043655396, + "learning_rate": 1.5770833333333335e-05, + "loss": 0.0861, + "step": 2130 + }, + { + "epoch": 21.88558644934484, + "grad_norm": 0.6466374397277832, + "learning_rate": 1.575e-05, + "loss": 0.0873, + "step": 2140 + }, + { + "epoch": 21.987855544902526, + "grad_norm": 2.8282721042633057, + "learning_rate": 1.5733333333333334e-05, + "loss": 0.0888, + "step": 2150 + }, + { + "epoch": 22.09012464046021, + "grad_norm": 0.5769690275192261, + "learning_rate": 1.5712500000000002e-05, + "loss": 0.0748, + "step": 2160 + }, + { + "epoch": 22.192393736017898, + "grad_norm": 0.5819457173347473, + "learning_rate": 1.5691666666666667e-05, + "loss": 0.0737, + "step": 2170 + }, + { + "epoch": 22.294662831575582, + "grad_norm": 0.6134530305862427, + "learning_rate": 1.5670833333333336e-05, + "loss": 0.0739, + "step": 2180 + }, + { + "epoch": 22.39693192713327, + "grad_norm": 0.6075708866119385, + "learning_rate": 1.565e-05, + "loss": 0.0742, + "step": 2190 + }, + { + "epoch": 22.499201022690954, + "grad_norm": 0.5924075245857239, + "learning_rate": 1.562916666666667e-05, + "loss": 0.0751, + "step": 2200 + }, + { + "epoch": 22.499201022690954, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.829688549041748, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8883, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.747, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.968, + "step": 2200 + }, + { + "epoch": 22.601470118248642, + "grad_norm": 0.6079824566841125, + "learning_rate": 1.5608333333333334e-05, + "loss": 0.0762, + "step": 2210 + }, + { + "epoch": 22.70373921380633, + "grad_norm": 0.6125743389129639, + "learning_rate": 1.5587500000000003e-05, + "loss": 0.0764, + "step": 2220 + }, + { + "epoch": 22.806008309364014, + "grad_norm": 0.5798956751823425, + "learning_rate": 1.5566666666666668e-05, + "loss": 0.0764, + "step": 2230 + }, + { + "epoch": 22.9082774049217, + "grad_norm": 0.6115660667419434, + "learning_rate": 1.5545833333333336e-05, + "loss": 0.0782, + "step": 2240 + }, + { + "epoch": 23.010546500479386, + "grad_norm": 0.5898561477661133, + "learning_rate": 1.5525e-05, + "loss": 0.0785, + "step": 2250 + }, + { + "epoch": 23.112815596037073, + "grad_norm": 0.5677723288536072, + "learning_rate": 1.550416666666667e-05, + "loss": 0.066, + "step": 2260 + }, + { + "epoch": 23.215084691594758, + "grad_norm": 0.5659050345420837, + "learning_rate": 1.5483333333333335e-05, + "loss": 0.0659, + "step": 2270 + }, + { + "epoch": 23.317353787152445, + "grad_norm": 0.5879548192024231, + "learning_rate": 1.54625e-05, + "loss": 0.0671, + "step": 2280 + }, + { + "epoch": 23.41962288271013, + "grad_norm": 0.5900695323944092, + "learning_rate": 1.5441666666666668e-05, + "loss": 0.0668, + "step": 2290 + }, + { + "epoch": 23.521891978267817, + "grad_norm": 0.5663672685623169, + "learning_rate": 1.5420833333333333e-05, + "loss": 0.0676, + "step": 2300 + }, + { + "epoch": 23.521891978267817, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.8641977310180664, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8705, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.829, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.979, + "step": 2300 + }, + { + "epoch": 23.624161073825505, + "grad_norm": 0.6123341917991638, + "learning_rate": 1.54e-05, + "loss": 0.0686, + "step": 2310 + }, + { + "epoch": 23.72643016938319, + "grad_norm": 0.5857561826705933, + "learning_rate": 1.5379166666666667e-05, + "loss": 0.0688, + "step": 2320 + }, + { + "epoch": 23.828699264940877, + "grad_norm": 0.6261973977088928, + "learning_rate": 1.5358333333333335e-05, + "loss": 0.0694, + "step": 2330 + }, + { + "epoch": 23.93096836049856, + "grad_norm": 0.5833300948143005, + "learning_rate": 1.53375e-05, + "loss": 0.0706, + "step": 2340 + }, + { + "epoch": 24.03323745605625, + "grad_norm": 0.5474048256874084, + "learning_rate": 1.531666666666667e-05, + "loss": 0.0676, + "step": 2350 + }, + { + "epoch": 24.135506551613933, + "grad_norm": 0.5484219193458557, + "learning_rate": 1.5295833333333334e-05, + "loss": 0.06, + "step": 2360 + }, + { + "epoch": 24.23777564717162, + "grad_norm": 0.5514199137687683, + "learning_rate": 1.5275000000000002e-05, + "loss": 0.0608, + "step": 2370 + }, + { + "epoch": 24.340044742729308, + "grad_norm": 0.5593263506889343, + "learning_rate": 1.5254166666666667e-05, + "loss": 0.0607, + "step": 2380 + }, + { + "epoch": 24.442313838286992, + "grad_norm": 0.5621811151504517, + "learning_rate": 1.5233333333333335e-05, + "loss": 0.0612, + "step": 2390 + }, + { + "epoch": 24.54458293384468, + "grad_norm": 0.5351541042327881, + "learning_rate": 1.52125e-05, + "loss": 0.0612, + "step": 2400 + }, + { + "epoch": 24.54458293384468, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.8911913633346558, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8888, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.745, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.968, + "step": 2400 + }, + { + "epoch": 24.646852029402364, + "grad_norm": 0.6047305464744568, + "learning_rate": 1.5191666666666669e-05, + "loss": 0.0628, + "step": 2410 + }, + { + "epoch": 24.749121124960052, + "grad_norm": 0.5636226534843445, + "learning_rate": 1.5170833333333334e-05, + "loss": 0.0624, + "step": 2420 + }, + { + "epoch": 24.851390220517736, + "grad_norm": 0.5656375288963318, + "learning_rate": 1.515e-05, + "loss": 0.0635, + "step": 2430 + }, + { + "epoch": 24.953659316075424, + "grad_norm": 0.5728236436843872, + "learning_rate": 1.5129166666666667e-05, + "loss": 0.0642, + "step": 2440 + }, + { + "epoch": 25.05592841163311, + "grad_norm": 0.5383201837539673, + "learning_rate": 1.5108333333333334e-05, + "loss": 0.0585, + "step": 2450 + }, + { + "epoch": 25.158197507190796, + "grad_norm": 0.5213799476623535, + "learning_rate": 1.50875e-05, + "loss": 0.0539, + "step": 2460 + }, + { + "epoch": 25.260466602748483, + "grad_norm": 0.5278561115264893, + "learning_rate": 1.5066666666666668e-05, + "loss": 0.0557, + "step": 2470 + }, + { + "epoch": 25.362735698306167, + "grad_norm": 0.5517110824584961, + "learning_rate": 1.5045833333333334e-05, + "loss": 0.054, + "step": 2480 + }, + { + "epoch": 25.465004793863855, + "grad_norm": 0.5266678333282471, + "learning_rate": 1.5025000000000001e-05, + "loss": 0.055, + "step": 2490 + }, + { + "epoch": 25.56727388942154, + "grad_norm": 0.5280548334121704, + "learning_rate": 1.5004166666666668e-05, + "loss": 0.056, + "step": 2500 + }, + { + "epoch": 25.56727388942154, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.9079244136810303, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8724, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.82, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.978, + "step": 2500 + }, + { + "epoch": 25.669542984979227, + "grad_norm": 0.543380081653595, + "learning_rate": 1.4983333333333334e-05, + "loss": 0.0567, + "step": 2510 + }, + { + "epoch": 25.77181208053691, + "grad_norm": 0.5638805031776428, + "learning_rate": 1.4962500000000003e-05, + "loss": 0.0576, + "step": 2520 + }, + { + "epoch": 25.8740811760946, + "grad_norm": 0.5791529417037964, + "learning_rate": 1.4941666666666668e-05, + "loss": 0.0585, + "step": 2530 + }, + { + "epoch": 25.976350271652286, + "grad_norm": 0.5550780892372131, + "learning_rate": 1.4920833333333336e-05, + "loss": 0.0583, + "step": 2540 + }, + { + "epoch": 26.07861936720997, + "grad_norm": 0.5031822323799133, + "learning_rate": 1.4900000000000001e-05, + "loss": 0.0521, + "step": 2550 + }, + { + "epoch": 26.18088846276766, + "grad_norm": 0.5042924880981445, + "learning_rate": 1.487916666666667e-05, + "loss": 0.0495, + "step": 2560 + }, + { + "epoch": 26.283157558325343, + "grad_norm": 0.5483749508857727, + "learning_rate": 1.4858333333333335e-05, + "loss": 0.0504, + "step": 2570 + }, + { + "epoch": 26.38542665388303, + "grad_norm": 0.508752703666687, + "learning_rate": 1.48375e-05, + "loss": 0.0509, + "step": 2580 + }, + { + "epoch": 26.487695749440714, + "grad_norm": 0.4737156331539154, + "learning_rate": 1.4816666666666668e-05, + "loss": 0.0507, + "step": 2590 + }, + { + "epoch": 26.589964844998402, + "grad_norm": 0.5402019023895264, + "learning_rate": 1.4795833333333333e-05, + "loss": 0.0513, + "step": 2600 + }, + { + "epoch": 26.589964844998402, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.9507412910461426, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8703, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.83, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.979, + "step": 2600 + }, + { + "epoch": 26.69223394055609, + "grad_norm": 0.5165488123893738, + "learning_rate": 1.4775000000000002e-05, + "loss": 0.0519, + "step": 2610 + }, + { + "epoch": 26.794503036113774, + "grad_norm": 0.5218144059181213, + "learning_rate": 1.4754166666666667e-05, + "loss": 0.0537, + "step": 2620 + }, + { + "epoch": 26.89677213167146, + "grad_norm": 0.5503118634223938, + "learning_rate": 1.4733333333333335e-05, + "loss": 0.053, + "step": 2630 + }, + { + "epoch": 26.999041227229146, + "grad_norm": 0.5653933882713318, + "learning_rate": 1.47125e-05, + "loss": 0.0533, + "step": 2640 + }, + { + "epoch": 27.101310322786833, + "grad_norm": 0.5077876448631287, + "learning_rate": 1.4691666666666669e-05, + "loss": 0.0473, + "step": 2650 + }, + { + "epoch": 27.203579418344518, + "grad_norm": 0.5198752880096436, + "learning_rate": 1.4670833333333334e-05, + "loss": 0.0472, + "step": 2660 + }, + { + "epoch": 27.305848513902205, + "grad_norm": 0.49260616302490234, + "learning_rate": 1.4650000000000002e-05, + "loss": 0.047, + "step": 2670 + }, + { + "epoch": 27.408117609459893, + "grad_norm": 0.5261008739471436, + "learning_rate": 1.4629166666666667e-05, + "loss": 0.0468, + "step": 2680 + }, + { + "epoch": 27.510386705017577, + "grad_norm": 0.5464609265327454, + "learning_rate": 1.4608333333333335e-05, + "loss": 0.0476, + "step": 2690 + }, + { + "epoch": 27.612655800575265, + "grad_norm": 0.526250422000885, + "learning_rate": 1.45875e-05, + "loss": 0.0488, + "step": 2700 + }, + { + "epoch": 27.612655800575265, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.963590145111084, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8709, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.827, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.978, + "step": 2700 + }, + { + "epoch": 27.71492489613295, + "grad_norm": 0.49739035964012146, + "learning_rate": 1.4566666666666669e-05, + "loss": 0.0478, + "step": 2710 + }, + { + "epoch": 27.817193991690637, + "grad_norm": 0.5307214260101318, + "learning_rate": 1.4545833333333334e-05, + "loss": 0.048, + "step": 2720 + }, + { + "epoch": 27.91946308724832, + "grad_norm": 0.5277805328369141, + "learning_rate": 1.4525e-05, + "loss": 0.0489, + "step": 2730 + }, + { + "epoch": 28.02173218280601, + "grad_norm": 0.490689218044281, + "learning_rate": 1.4504166666666667e-05, + "loss": 0.0488, + "step": 2740 + }, + { + "epoch": 28.124001278363693, + "grad_norm": 0.5121079683303833, + "learning_rate": 1.4483333333333334e-05, + "loss": 0.0426, + "step": 2750 + }, + { + "epoch": 28.22627037392138, + "grad_norm": 0.5197770595550537, + "learning_rate": 1.4462500000000001e-05, + "loss": 0.0433, + "step": 2760 + }, + { + "epoch": 28.328539469479068, + "grad_norm": 0.5023459196090698, + "learning_rate": 1.4441666666666668e-05, + "loss": 0.0428, + "step": 2770 + }, + { + "epoch": 28.430808565036752, + "grad_norm": 0.47794410586357117, + "learning_rate": 1.4420833333333334e-05, + "loss": 0.0431, + "step": 2780 + }, + { + "epoch": 28.53307766059444, + "grad_norm": 0.4780057370662689, + "learning_rate": 1.4400000000000001e-05, + "loss": 0.0439, + "step": 2790 + }, + { + "epoch": 28.635346756152124, + "grad_norm": 0.5096541047096252, + "learning_rate": 1.4379166666666668e-05, + "loss": 0.0443, + "step": 2800 + }, + { + "epoch": 28.635346756152124, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.9908784627914429, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8556, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.897, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.987, + "step": 2800 + }, + { + "epoch": 28.737615851709812, + "grad_norm": 0.49939730763435364, + "learning_rate": 1.4358333333333334e-05, + "loss": 0.045, + "step": 2810 + }, + { + "epoch": 28.839884947267496, + "grad_norm": 0.5106296539306641, + "learning_rate": 1.4337500000000001e-05, + "loss": 0.045, + "step": 2820 + }, + { + "epoch": 28.942154042825184, + "grad_norm": 0.4849531054496765, + "learning_rate": 1.4316666666666668e-05, + "loss": 0.0451, + "step": 2830 + }, + { + "epoch": 29.04442313838287, + "grad_norm": 0.46746712923049927, + "learning_rate": 1.4295833333333335e-05, + "loss": 0.0429, + "step": 2840 + }, + { + "epoch": 29.146692233940556, + "grad_norm": 0.4314974546432495, + "learning_rate": 1.4275000000000001e-05, + "loss": 0.0397, + "step": 2850 + }, + { + "epoch": 29.248961329498243, + "grad_norm": 0.4507407546043396, + "learning_rate": 1.4254166666666668e-05, + "loss": 0.0394, + "step": 2860 + }, + { + "epoch": 29.351230425055927, + "grad_norm": 0.4565179646015167, + "learning_rate": 1.4233333333333335e-05, + "loss": 0.0402, + "step": 2870 + }, + { + "epoch": 29.453499520613615, + "grad_norm": 0.4820208251476288, + "learning_rate": 1.42125e-05, + "loss": 0.0403, + "step": 2880 + }, + { + "epoch": 29.5557686161713, + "grad_norm": 0.47718533873558044, + "learning_rate": 1.4191666666666668e-05, + "loss": 0.0404, + "step": 2890 + }, + { + "epoch": 29.658037711728987, + "grad_norm": 0.5102068185806274, + "learning_rate": 1.4170833333333333e-05, + "loss": 0.0418, + "step": 2900 + }, + { + "epoch": 29.658037711728987, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.013613700866699, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8531, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.909, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.989, + "step": 2900 + }, + { + "epoch": 29.760306807286675, + "grad_norm": 0.49850982427597046, + "learning_rate": 1.4150000000000002e-05, + "loss": 0.041, + "step": 2910 + }, + { + "epoch": 29.86257590284436, + "grad_norm": 0.5028281211853027, + "learning_rate": 1.4129166666666667e-05, + "loss": 0.0419, + "step": 2920 + }, + { + "epoch": 29.964844998402047, + "grad_norm": 0.4753844141960144, + "learning_rate": 1.4108333333333335e-05, + "loss": 0.0421, + "step": 2930 + }, + { + "epoch": 30.06711409395973, + "grad_norm": 0.45387542247772217, + "learning_rate": 1.40875e-05, + "loss": 0.0381, + "step": 2940 + }, + { + "epoch": 30.16938318951742, + "grad_norm": 0.4576801359653473, + "learning_rate": 1.4066666666666669e-05, + "loss": 0.0365, + "step": 2950 + }, + { + "epoch": 30.271652285075103, + "grad_norm": 1.5916332006454468, + "learning_rate": 1.4045833333333334e-05, + "loss": 0.0371, + "step": 2960 + }, + { + "epoch": 30.37392138063279, + "grad_norm": 28.43450164794922, + "learning_rate": 1.4029166666666668e-05, + "loss": 0.0403, + "step": 2970 + }, + { + "epoch": 30.476190476190474, + "grad_norm": 23.308544158935547, + "learning_rate": 1.4010416666666669e-05, + "loss": 0.3001, + "step": 2980 + }, + { + "epoch": 30.578459571748162, + "grad_norm": 0.9334068298339844, + "learning_rate": 1.3989583333333334e-05, + "loss": 0.2114, + "step": 2990 + }, + { + "epoch": 30.68072866730585, + "grad_norm": 0.5433509945869446, + "learning_rate": 1.3968750000000002e-05, + "loss": 0.0448, + "step": 3000 + }, + { + "epoch": 30.68072866730585, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.014845609664917, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8553, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.899, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.987, + "step": 3000 + }, + { + "epoch": 30.782997762863534, + "grad_norm": 0.5084338188171387, + "learning_rate": 1.3947916666666667e-05, + "loss": 0.042, + "step": 3010 + }, + { + "epoch": 30.88526685842122, + "grad_norm": 0.4836946725845337, + "learning_rate": 1.3927083333333336e-05, + "loss": 0.0396, + "step": 3020 + }, + { + "epoch": 30.987535953978906, + "grad_norm": 0.4835638701915741, + "learning_rate": 1.3906250000000001e-05, + "loss": 0.0397, + "step": 3030 + }, + { + "epoch": 31.089805049536594, + "grad_norm": 0.4484618008136749, + "learning_rate": 1.3885416666666666e-05, + "loss": 0.0359, + "step": 3040 + }, + { + "epoch": 31.192074145094278, + "grad_norm": 0.4576917886734009, + "learning_rate": 1.3864583333333334e-05, + "loss": 0.0352, + "step": 3050 + }, + { + "epoch": 31.294343240651965, + "grad_norm": 0.4618643820285797, + "learning_rate": 1.3843750000000001e-05, + "loss": 0.0351, + "step": 3060 + }, + { + "epoch": 31.396612336209653, + "grad_norm": 0.4532334804534912, + "learning_rate": 1.3822916666666668e-05, + "loss": 0.0353, + "step": 3070 + }, + { + "epoch": 31.498881431767337, + "grad_norm": 0.43523749709129333, + "learning_rate": 1.3802083333333335e-05, + "loss": 0.0357, + "step": 3080 + }, + { + "epoch": 31.601150527325025, + "grad_norm": 0.4621034860610962, + "learning_rate": 1.3781250000000001e-05, + "loss": 0.036, + "step": 3090 + }, + { + "epoch": 31.70341962288271, + "grad_norm": 0.4407444894313812, + "learning_rate": 1.3760416666666668e-05, + "loss": 0.0365, + "step": 3100 + }, + { + "epoch": 31.70341962288271, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.0446879863739014, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8565, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.893, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.987, + "step": 3100 + }, + { + "epoch": 31.805688718440397, + "grad_norm": 0.46584275364875793, + "learning_rate": 1.3739583333333335e-05, + "loss": 0.0367, + "step": 3110 + }, + { + "epoch": 31.90795781399808, + "grad_norm": 0.4602925181388855, + "learning_rate": 1.3718750000000001e-05, + "loss": 0.0368, + "step": 3120 + }, + { + "epoch": 32.01022690955577, + "grad_norm": 0.4346022605895996, + "learning_rate": 1.3697916666666668e-05, + "loss": 0.0398, + "step": 3130 + }, + { + "epoch": 32.11249600511346, + "grad_norm": 0.42625728249549866, + "learning_rate": 1.3677083333333335e-05, + "loss": 0.0327, + "step": 3140 + }, + { + "epoch": 32.214765100671144, + "grad_norm": 0.4581526219844818, + "learning_rate": 1.3656250000000002e-05, + "loss": 0.0325, + "step": 3150 + }, + { + "epoch": 32.317034196228825, + "grad_norm": 0.41788893938064575, + "learning_rate": 1.3635416666666668e-05, + "loss": 0.0323, + "step": 3160 + }, + { + "epoch": 32.41930329178651, + "grad_norm": 0.39856305718421936, + "learning_rate": 1.3614583333333335e-05, + "loss": 0.0331, + "step": 3170 + }, + { + "epoch": 32.5215723873442, + "grad_norm": 0.4417785704135895, + "learning_rate": 1.3593750000000002e-05, + "loss": 0.0336, + "step": 3180 + }, + { + "epoch": 32.62384148290189, + "grad_norm": 0.45321398973464966, + "learning_rate": 1.3572916666666667e-05, + "loss": 0.0339, + "step": 3190 + }, + { + "epoch": 32.72611057845957, + "grad_norm": 0.43776148557662964, + "learning_rate": 1.3552083333333335e-05, + "loss": 0.0337, + "step": 3200 + }, + { + "epoch": 32.72611057845957, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.0593974590301514, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8377, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.98, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.997, + "step": 3200 + }, + { + "epoch": 32.828379674017256, + "grad_norm": 0.4577961564064026, + "learning_rate": 1.353125e-05, + "loss": 0.0342, + "step": 3210 + }, + { + "epoch": 32.930648769574944, + "grad_norm": 0.43651047348976135, + "learning_rate": 1.3510416666666669e-05, + "loss": 0.0352, + "step": 3220 + }, + { + "epoch": 33.03291786513263, + "grad_norm": 0.4300122559070587, + "learning_rate": 1.3489583333333334e-05, + "loss": 0.0331, + "step": 3230 + }, + { + "epoch": 33.13518696069032, + "grad_norm": 0.43906641006469727, + "learning_rate": 1.3468750000000002e-05, + "loss": 0.0304, + "step": 3240 + }, + { + "epoch": 33.237456056248, + "grad_norm": 0.40852677822113037, + "learning_rate": 1.3447916666666667e-05, + "loss": 0.0304, + "step": 3250 + }, + { + "epoch": 33.33972515180569, + "grad_norm": 0.42338284850120544, + "learning_rate": 1.3427083333333336e-05, + "loss": 0.0307, + "step": 3260 + }, + { + "epoch": 33.441994247363375, + "grad_norm": 0.4441679120063782, + "learning_rate": 1.340625e-05, + "loss": 0.031, + "step": 3270 + }, + { + "epoch": 33.54426334292106, + "grad_norm": 0.431755393743515, + "learning_rate": 1.3385416666666669e-05, + "loss": 0.031, + "step": 3280 + }, + { + "epoch": 33.646532438478744, + "grad_norm": 0.42783161997795105, + "learning_rate": 1.3364583333333334e-05, + "loss": 0.0316, + "step": 3290 + }, + { + "epoch": 33.74880153403643, + "grad_norm": 0.4471502900123596, + "learning_rate": 1.3343750000000002e-05, + "loss": 0.0318, + "step": 3300 + }, + { + "epoch": 33.74880153403643, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.0793018341064453, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.9135, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.632, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.954, + "step": 3300 + }, + { + "epoch": 33.85107062959412, + "grad_norm": 0.42543718218803406, + "learning_rate": 1.3322916666666668e-05, + "loss": 0.0317, + "step": 3310 + }, + { + "epoch": 33.95333972515181, + "grad_norm": 0.4455837905406952, + "learning_rate": 1.3302083333333336e-05, + "loss": 0.0321, + "step": 3320 + }, + { + "epoch": 34.055608820709494, + "grad_norm": 0.451027512550354, + "learning_rate": 1.3281250000000001e-05, + "loss": 0.0297, + "step": 3330 + }, + { + "epoch": 34.157877916267175, + "grad_norm": 0.43288454413414, + "learning_rate": 1.3260416666666666e-05, + "loss": 0.0285, + "step": 3340 + }, + { + "epoch": 34.26014701182486, + "grad_norm": 0.41506046056747437, + "learning_rate": 1.3239583333333334e-05, + "loss": 0.0285, + "step": 3350 + }, + { + "epoch": 34.36241610738255, + "grad_norm": 0.4198153614997864, + "learning_rate": 1.321875e-05, + "loss": 0.0287, + "step": 3360 + }, + { + "epoch": 34.46468520294024, + "grad_norm": 0.4576322138309479, + "learning_rate": 1.3197916666666668e-05, + "loss": 0.0293, + "step": 3370 + }, + { + "epoch": 34.566954298497926, + "grad_norm": 0.4268178343772888, + "learning_rate": 1.3177083333333333e-05, + "loss": 0.0297, + "step": 3380 + }, + { + "epoch": 34.66922339405561, + "grad_norm": 0.4406622648239136, + "learning_rate": 1.3156250000000001e-05, + "loss": 0.0298, + "step": 3390 + }, + { + "epoch": 34.771492489613294, + "grad_norm": 0.44503384828567505, + "learning_rate": 1.3135416666666666e-05, + "loss": 0.0302, + "step": 3400 + }, + { + "epoch": 34.771492489613294, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1051576137542725, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7904, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.198, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.025, + "step": 3400 + }, + { + "epoch": 34.87376158517098, + "grad_norm": 0.4410327672958374, + "learning_rate": 1.3114583333333335e-05, + "loss": 0.0308, + "step": 3410 + }, + { + "epoch": 34.97603068072867, + "grad_norm": 0.4527595639228821, + "learning_rate": 1.309375e-05, + "loss": 0.0307, + "step": 3420 + }, + { + "epoch": 35.07829977628635, + "grad_norm": 0.3983278274536133, + "learning_rate": 1.3072916666666668e-05, + "loss": 0.0275, + "step": 3430 + }, + { + "epoch": 35.18056887184404, + "grad_norm": 0.405274361371994, + "learning_rate": 1.3052083333333335e-05, + "loss": 0.027, + "step": 3440 + }, + { + "epoch": 35.282837967401726, + "grad_norm": 0.4225080907344818, + "learning_rate": 1.3031250000000002e-05, + "loss": 0.0276, + "step": 3450 + }, + { + "epoch": 35.38510706295941, + "grad_norm": 0.3988070487976074, + "learning_rate": 1.3010416666666668e-05, + "loss": 0.0271, + "step": 3460 + }, + { + "epoch": 35.4873761585171, + "grad_norm": 0.613850474357605, + "learning_rate": 1.2989583333333335e-05, + "loss": 0.0271, + "step": 3470 + }, + { + "epoch": 35.58964525407478, + "grad_norm": 0.42930155992507935, + "learning_rate": 1.2968750000000002e-05, + "loss": 0.0269, + "step": 3480 + }, + { + "epoch": 35.69191434963247, + "grad_norm": 0.4209059178829193, + "learning_rate": 1.2947916666666667e-05, + "loss": 0.0281, + "step": 3490 + }, + { + "epoch": 35.79418344519016, + "grad_norm": 0.4427293837070465, + "learning_rate": 1.2927083333333335e-05, + "loss": 0.0282, + "step": 3500 + }, + { + "epoch": 35.79418344519016, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1160192489624023, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8212, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.056, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.007, + "step": 3500 + }, + { + "epoch": 35.896452540747845, + "grad_norm": 0.42366185784339905, + "learning_rate": 1.290625e-05, + "loss": 0.0289, + "step": 3510 + }, + { + "epoch": 35.99872163630553, + "grad_norm": 0.4120742380619049, + "learning_rate": 1.2885416666666669e-05, + "loss": 0.0287, + "step": 3520 + }, + { + "epoch": 36.10099073186321, + "grad_norm": 0.38890495896339417, + "learning_rate": 1.2864583333333334e-05, + "loss": 0.0258, + "step": 3530 + }, + { + "epoch": 36.2032598274209, + "grad_norm": 0.4076452851295471, + "learning_rate": 1.2843750000000002e-05, + "loss": 0.0258, + "step": 3540 + }, + { + "epoch": 36.30552892297859, + "grad_norm": 0.3809320032596588, + "learning_rate": 1.2822916666666667e-05, + "loss": 0.0255, + "step": 3550 + }, + { + "epoch": 36.407798018536276, + "grad_norm": 0.39650237560272217, + "learning_rate": 1.2802083333333336e-05, + "loss": 0.026, + "step": 3560 + }, + { + "epoch": 36.51006711409396, + "grad_norm": 0.4009132385253906, + "learning_rate": 1.278125e-05, + "loss": 0.0265, + "step": 3570 + }, + { + "epoch": 36.612336209651644, + "grad_norm": 0.42308393120765686, + "learning_rate": 1.2760416666666669e-05, + "loss": 0.0263, + "step": 3580 + }, + { + "epoch": 36.71460530520933, + "grad_norm": 0.4044691324234009, + "learning_rate": 1.2739583333333334e-05, + "loss": 0.0265, + "step": 3590 + }, + { + "epoch": 36.81687440076702, + "grad_norm": 0.4140937030315399, + "learning_rate": 1.2718750000000003e-05, + "loss": 0.0266, + "step": 3600 + }, + { + "epoch": 36.81687440076702, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.13891339302063, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8236, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.045, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.006, + "step": 3600 + }, + { + "epoch": 36.91914349632471, + "grad_norm": 0.4160206615924835, + "learning_rate": 1.2697916666666668e-05, + "loss": 0.0269, + "step": 3610 + }, + { + "epoch": 37.02141259188239, + "grad_norm": 0.39423078298568726, + "learning_rate": 1.2677083333333336e-05, + "loss": 0.026, + "step": 3620 + }, + { + "epoch": 37.123681687440076, + "grad_norm": 0.3859294056892395, + "learning_rate": 1.2656250000000001e-05, + "loss": 0.0238, + "step": 3630 + }, + { + "epoch": 37.225950782997764, + "grad_norm": 0.40413331985473633, + "learning_rate": 1.2635416666666666e-05, + "loss": 0.0244, + "step": 3640 + }, + { + "epoch": 37.32821987855545, + "grad_norm": 0.39002159237861633, + "learning_rate": 1.2614583333333334e-05, + "loss": 0.0244, + "step": 3650 + }, + { + "epoch": 37.43048897411313, + "grad_norm": 0.403145432472229, + "learning_rate": 1.259375e-05, + "loss": 0.0245, + "step": 3660 + }, + { + "epoch": 37.53275806967082, + "grad_norm": 0.42878827452659607, + "learning_rate": 1.2572916666666668e-05, + "loss": 0.0249, + "step": 3670 + }, + { + "epoch": 37.63502716522851, + "grad_norm": 0.4047834277153015, + "learning_rate": 1.2552083333333333e-05, + "loss": 0.0246, + "step": 3680 + }, + { + "epoch": 37.737296260786195, + "grad_norm": 0.4242531955242157, + "learning_rate": 1.2531250000000001e-05, + "loss": 0.0254, + "step": 3690 + }, + { + "epoch": 37.83956535634388, + "grad_norm": 0.3883196711540222, + "learning_rate": 1.2510416666666666e-05, + "loss": 0.025, + "step": 3700 + }, + { + "epoch": 37.83956535634388, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.135497570037842, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.828, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.024, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.003, + "step": 3700 + }, + { + "epoch": 37.94183445190156, + "grad_norm": 0.40664049983024597, + "learning_rate": 1.2489583333333335e-05, + "loss": 0.0249, + "step": 3710 + }, + { + "epoch": 38.04410354745925, + "grad_norm": 0.37315633893013, + "learning_rate": 1.246875e-05, + "loss": 0.0242, + "step": 3720 + }, + { + "epoch": 38.14637264301694, + "grad_norm": 0.37339115142822266, + "learning_rate": 1.2447916666666668e-05, + "loss": 0.0229, + "step": 3730 + }, + { + "epoch": 38.24864173857463, + "grad_norm": 0.38532310724258423, + "learning_rate": 1.2427083333333333e-05, + "loss": 0.0227, + "step": 3740 + }, + { + "epoch": 38.350910834132314, + "grad_norm": 0.3730473220348358, + "learning_rate": 1.2406250000000002e-05, + "loss": 0.0227, + "step": 3750 + }, + { + "epoch": 38.453179929689995, + "grad_norm": 0.37635692954063416, + "learning_rate": 1.2385416666666667e-05, + "loss": 0.0234, + "step": 3760 + }, + { + "epoch": 38.55544902524768, + "grad_norm": 0.38997572660446167, + "learning_rate": 1.2364583333333335e-05, + "loss": 0.0233, + "step": 3770 + }, + { + "epoch": 38.65771812080537, + "grad_norm": 0.38781270384788513, + "learning_rate": 1.234375e-05, + "loss": 0.0237, + "step": 3780 + }, + { + "epoch": 38.75998721636306, + "grad_norm": 0.42525115609169006, + "learning_rate": 1.2322916666666667e-05, + "loss": 0.0237, + "step": 3790 + }, + { + "epoch": 38.86225631192074, + "grad_norm": 0.3896453380584717, + "learning_rate": 1.2302083333333335e-05, + "loss": 0.0239, + "step": 3800 + }, + { + "epoch": 38.86225631192074, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.166919231414795, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8189, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.067, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.008, + "step": 3800 + }, + { + "epoch": 38.964525407478426, + "grad_norm": 0.3976113498210907, + "learning_rate": 1.228125e-05, + "loss": 0.0242, + "step": 3810 + }, + { + "epoch": 39.066794503036114, + "grad_norm": 0.37402892112731934, + "learning_rate": 1.2260416666666669e-05, + "loss": 0.0228, + "step": 3820 + }, + { + "epoch": 39.1690635985938, + "grad_norm": 0.4018329679965973, + "learning_rate": 1.2239583333333334e-05, + "loss": 0.0219, + "step": 3830 + }, + { + "epoch": 39.27133269415149, + "grad_norm": 0.3905806839466095, + "learning_rate": 1.2218750000000002e-05, + "loss": 0.0219, + "step": 3840 + }, + { + "epoch": 39.37360178970917, + "grad_norm": 0.39368128776550293, + "learning_rate": 1.2197916666666667e-05, + "loss": 0.0223, + "step": 3850 + }, + { + "epoch": 39.47587088526686, + "grad_norm": 0.40275242924690247, + "learning_rate": 1.2177083333333336e-05, + "loss": 0.0225, + "step": 3860 + }, + { + "epoch": 39.578139980824545, + "grad_norm": 0.377655953168869, + "learning_rate": 1.215625e-05, + "loss": 0.0223, + "step": 3870 + }, + { + "epoch": 39.68040907638223, + "grad_norm": 0.35910850763320923, + "learning_rate": 1.2135416666666669e-05, + "loss": 0.0222, + "step": 3880 + }, + { + "epoch": 39.782678171939914, + "grad_norm": 0.40253758430480957, + "learning_rate": 1.2114583333333334e-05, + "loss": 0.0229, + "step": 3890 + }, + { + "epoch": 39.8849472674976, + "grad_norm": 0.4119781255722046, + "learning_rate": 1.2093750000000003e-05, + "loss": 0.0231, + "step": 3900 + }, + { + "epoch": 39.8849472674976, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1834068298339844, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8776, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.796, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.975, + "step": 3900 + }, + { + "epoch": 39.98721636305529, + "grad_norm": 0.40584734082221985, + "learning_rate": 1.2072916666666668e-05, + "loss": 0.0229, + "step": 3910 + }, + { + "epoch": 40.08948545861298, + "grad_norm": 0.3695628345012665, + "learning_rate": 1.2052083333333336e-05, + "loss": 0.0207, + "step": 3920 + }, + { + "epoch": 40.191754554170664, + "grad_norm": 0.38678476214408875, + "learning_rate": 1.2031250000000001e-05, + "loss": 0.0205, + "step": 3930 + }, + { + "epoch": 40.294023649728345, + "grad_norm": 0.349589079618454, + "learning_rate": 1.2010416666666666e-05, + "loss": 0.0208, + "step": 3940 + }, + { + "epoch": 40.39629274528603, + "grad_norm": 0.3602575659751892, + "learning_rate": 1.1989583333333335e-05, + "loss": 0.0207, + "step": 3950 + }, + { + "epoch": 40.49856184084372, + "grad_norm": 0.34940171241760254, + "learning_rate": 1.196875e-05, + "loss": 0.0208, + "step": 3960 + }, + { + "epoch": 40.60083093640141, + "grad_norm": 0.42674344778060913, + "learning_rate": 1.1947916666666668e-05, + "loss": 0.0214, + "step": 3970 + }, + { + "epoch": 40.703100031959096, + "grad_norm": 0.3709782361984253, + "learning_rate": 1.1927083333333333e-05, + "loss": 0.0215, + "step": 3980 + }, + { + "epoch": 40.80536912751678, + "grad_norm": 0.37002718448638916, + "learning_rate": 1.1906250000000001e-05, + "loss": 0.0215, + "step": 3990 + }, + { + "epoch": 40.907638223074464, + "grad_norm": 0.3808917999267578, + "learning_rate": 1.1885416666666666e-05, + "loss": 0.0216, + "step": 4000 + }, + { + "epoch": 40.907638223074464, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.173332452774048, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8356, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.99, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.999, + "step": 4000 + }, + { + "epoch": 41.00990731863215, + "grad_norm": 0.5391054153442383, + "learning_rate": 1.1864583333333335e-05, + "loss": 0.0214, + "step": 4010 + }, + { + "epoch": 41.11217641418984, + "grad_norm": 0.36861568689346313, + "learning_rate": 1.184375e-05, + "loss": 0.0203, + "step": 4020 + }, + { + "epoch": 41.21444550974752, + "grad_norm": 0.3627229630947113, + "learning_rate": 1.1822916666666668e-05, + "loss": 0.0203, + "step": 4030 + }, + { + "epoch": 41.31671460530521, + "grad_norm": 0.37450557947158813, + "learning_rate": 1.1802083333333333e-05, + "loss": 0.02, + "step": 4040 + }, + { + "epoch": 41.418983700862896, + "grad_norm": 0.3615454435348511, + "learning_rate": 1.1781250000000002e-05, + "loss": 0.0204, + "step": 4050 + }, + { + "epoch": 41.52125279642058, + "grad_norm": 0.3542068302631378, + "learning_rate": 1.1760416666666667e-05, + "loss": 0.0203, + "step": 4060 + }, + { + "epoch": 41.62352189197827, + "grad_norm": 0.3920552134513855, + "learning_rate": 1.1739583333333335e-05, + "loss": 0.0204, + "step": 4070 + }, + { + "epoch": 41.72579098753595, + "grad_norm": 0.3835306763648987, + "learning_rate": 1.171875e-05, + "loss": 0.0206, + "step": 4080 + }, + { + "epoch": 41.82806008309364, + "grad_norm": 0.35131242871284485, + "learning_rate": 1.1697916666666667e-05, + "loss": 0.0208, + "step": 4090 + }, + { + "epoch": 41.93032917865133, + "grad_norm": 0.3543274402618408, + "learning_rate": 1.1677083333333334e-05, + "loss": 0.0202, + "step": 4100 + }, + { + "epoch": 41.93032917865133, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1988096237182617, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8148, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.086, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.011, + "step": 4100 + }, + { + "epoch": 42.032598274209015, + "grad_norm": 0.33442866802215576, + "learning_rate": 1.165625e-05, + "loss": 0.0198, + "step": 4110 + }, + { + "epoch": 42.134867369766695, + "grad_norm": 0.3324146866798401, + "learning_rate": 1.1635416666666667e-05, + "loss": 0.0192, + "step": 4120 + }, + { + "epoch": 42.23713646532438, + "grad_norm": 0.3776349723339081, + "learning_rate": 1.1614583333333334e-05, + "loss": 0.0191, + "step": 4130 + }, + { + "epoch": 42.33940556088207, + "grad_norm": 0.3571579158306122, + "learning_rate": 1.159375e-05, + "loss": 0.0192, + "step": 4140 + }, + { + "epoch": 42.44167465643976, + "grad_norm": 0.37327298521995544, + "learning_rate": 1.1572916666666667e-05, + "loss": 0.0192, + "step": 4150 + }, + { + "epoch": 42.543943751997446, + "grad_norm": 0.3787640631198883, + "learning_rate": 1.1552083333333334e-05, + "loss": 0.0196, + "step": 4160 + }, + { + "epoch": 42.64621284755513, + "grad_norm": 0.36109158396720886, + "learning_rate": 1.153125e-05, + "loss": 0.0193, + "step": 4170 + }, + { + "epoch": 42.748481943112814, + "grad_norm": 0.34873998165130615, + "learning_rate": 1.151041666666667e-05, + "loss": 0.0194, + "step": 4180 + }, + { + "epoch": 42.8507510386705, + "grad_norm": 0.393928587436676, + "learning_rate": 1.1489583333333334e-05, + "loss": 0.0196, + "step": 4190 + }, + { + "epoch": 42.95302013422819, + "grad_norm": 0.3687518537044525, + "learning_rate": 1.1468750000000003e-05, + "loss": 0.0198, + "step": 4200 + }, + { + "epoch": 42.95302013422819, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2081563472747803, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7395, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.434, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.054, + "step": 4200 + }, + { + "epoch": 43.05528922978588, + "grad_norm": 0.3610316514968872, + "learning_rate": 1.1447916666666668e-05, + "loss": 0.0187, + "step": 4210 + }, + { + "epoch": 43.15755832534356, + "grad_norm": 0.3438417911529541, + "learning_rate": 1.1427083333333336e-05, + "loss": 0.0181, + "step": 4220 + }, + { + "epoch": 43.259827420901246, + "grad_norm": 0.33607029914855957, + "learning_rate": 1.1406250000000001e-05, + "loss": 0.018, + "step": 4230 + }, + { + "epoch": 43.362096516458934, + "grad_norm": 0.3655332326889038, + "learning_rate": 1.1385416666666666e-05, + "loss": 0.018, + "step": 4240 + }, + { + "epoch": 43.46436561201662, + "grad_norm": 0.3450184762477875, + "learning_rate": 1.1364583333333335e-05, + "loss": 0.0182, + "step": 4250 + }, + { + "epoch": 43.5666347075743, + "grad_norm": 0.34371188282966614, + "learning_rate": 1.134375e-05, + "loss": 0.0185, + "step": 4260 + }, + { + "epoch": 43.66890380313199, + "grad_norm": 0.3620694875717163, + "learning_rate": 1.1322916666666668e-05, + "loss": 0.0188, + "step": 4270 + }, + { + "epoch": 43.77117289868968, + "grad_norm": 0.34563344717025757, + "learning_rate": 1.1302083333333333e-05, + "loss": 0.0185, + "step": 4280 + }, + { + "epoch": 43.873441994247365, + "grad_norm": 0.3630271255970001, + "learning_rate": 1.1281250000000001e-05, + "loss": 0.0189, + "step": 4290 + }, + { + "epoch": 43.97571108980505, + "grad_norm": 0.3588694930076599, + "learning_rate": 1.1260416666666666e-05, + "loss": 0.0192, + "step": 4300 + }, + { + "epoch": 43.97571108980505, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.21140456199646, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7918, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.192, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.024, + "step": 4300 + }, + { + "epoch": 44.07798018536273, + "grad_norm": 0.3162562847137451, + "learning_rate": 1.1239583333333335e-05, + "loss": 0.0172, + "step": 4310 + }, + { + "epoch": 44.18024928092042, + "grad_norm": 0.366394966840744, + "learning_rate": 1.121875e-05, + "loss": 0.017, + "step": 4320 + }, + { + "epoch": 44.28251837647811, + "grad_norm": 0.35083746910095215, + "learning_rate": 1.1197916666666668e-05, + "loss": 0.0178, + "step": 4330 + }, + { + "epoch": 44.384787472035796, + "grad_norm": 0.35160747170448303, + "learning_rate": 1.1177083333333333e-05, + "loss": 0.0174, + "step": 4340 + }, + { + "epoch": 44.48705656759348, + "grad_norm": 0.35368478298187256, + "learning_rate": 1.1156250000000002e-05, + "loss": 0.0175, + "step": 4350 + }, + { + "epoch": 44.589325663151165, + "grad_norm": 0.3595126271247864, + "learning_rate": 1.1135416666666667e-05, + "loss": 0.0176, + "step": 4360 + }, + { + "epoch": 44.69159475870885, + "grad_norm": 0.34251803159713745, + "learning_rate": 1.1114583333333335e-05, + "loss": 0.0179, + "step": 4370 + }, + { + "epoch": 44.79386385426654, + "grad_norm": 0.3488711416721344, + "learning_rate": 1.109375e-05, + "loss": 0.018, + "step": 4380 + }, + { + "epoch": 44.89613294982423, + "grad_norm": 0.35304006934165955, + "learning_rate": 1.1072916666666667e-05, + "loss": 0.018, + "step": 4390 + }, + { + "epoch": 44.99840204538191, + "grad_norm": 0.3306570053100586, + "learning_rate": 1.1052083333333334e-05, + "loss": 0.018, + "step": 4400 + }, + { + "epoch": 44.99840204538191, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.226471424102783, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7401, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.432, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.054, + "step": 4400 + }, + { + "epoch": 45.100671140939596, + "grad_norm": 0.3414161801338196, + "learning_rate": 1.103125e-05, + "loss": 0.0163, + "step": 4410 + }, + { + "epoch": 45.202940236497284, + "grad_norm": 0.34503790736198425, + "learning_rate": 1.1010416666666667e-05, + "loss": 0.0167, + "step": 4420 + }, + { + "epoch": 45.30520933205497, + "grad_norm": 0.3578576147556305, + "learning_rate": 1.0989583333333334e-05, + "loss": 0.0165, + "step": 4430 + }, + { + "epoch": 45.40747842761266, + "grad_norm": 0.34795865416526794, + "learning_rate": 1.096875e-05, + "loss": 0.0165, + "step": 4440 + }, + { + "epoch": 45.50974752317034, + "grad_norm": 0.35079431533813477, + "learning_rate": 1.0947916666666667e-05, + "loss": 0.0168, + "step": 4450 + }, + { + "epoch": 45.61201661872803, + "grad_norm": 0.30789047479629517, + "learning_rate": 1.0927083333333334e-05, + "loss": 0.0164, + "step": 4460 + }, + { + "epoch": 45.714285714285715, + "grad_norm": 0.370721697807312, + "learning_rate": 1.090625e-05, + "loss": 0.0174, + "step": 4470 + }, + { + "epoch": 45.8165548098434, + "grad_norm": 0.34695321321487427, + "learning_rate": 1.0885416666666668e-05, + "loss": 0.0171, + "step": 4480 + }, + { + "epoch": 45.918823905401084, + "grad_norm": 0.3586573600769043, + "learning_rate": 1.0864583333333334e-05, + "loss": 0.0171, + "step": 4490 + }, + { + "epoch": 46.02109300095877, + "grad_norm": 0.3352271318435669, + "learning_rate": 1.0843750000000001e-05, + "loss": 0.0172, + "step": 4500 + }, + { + "epoch": 46.02109300095877, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2414653301239014, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.694, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.646, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.081, + "step": 4500 + }, + { + "epoch": 46.12336209651646, + "grad_norm": 0.3106061518192291, + "learning_rate": 1.0822916666666668e-05, + "loss": 0.0154, + "step": 4510 + }, + { + "epoch": 46.22563119207415, + "grad_norm": 0.3251485824584961, + "learning_rate": 1.0802083333333334e-05, + "loss": 0.0159, + "step": 4520 + }, + { + "epoch": 46.327900287631834, + "grad_norm": 0.3339298367500305, + "learning_rate": 1.0781250000000001e-05, + "loss": 0.0157, + "step": 4530 + }, + { + "epoch": 46.430169383189515, + "grad_norm": 0.33735695481300354, + "learning_rate": 1.0760416666666666e-05, + "loss": 0.0161, + "step": 4540 + }, + { + "epoch": 46.5324384787472, + "grad_norm": 0.3049178719520569, + "learning_rate": 1.0739583333333335e-05, + "loss": 0.0162, + "step": 4550 + }, + { + "epoch": 46.63470757430489, + "grad_norm": 0.3524739742279053, + "learning_rate": 1.071875e-05, + "loss": 0.016, + "step": 4560 + }, + { + "epoch": 46.73697666986258, + "grad_norm": 0.3321375250816345, + "learning_rate": 1.0697916666666668e-05, + "loss": 0.016, + "step": 4570 + }, + { + "epoch": 46.83924576542026, + "grad_norm": 0.3346993029117584, + "learning_rate": 1.0677083333333333e-05, + "loss": 0.0165, + "step": 4580 + }, + { + "epoch": 46.94151486097795, + "grad_norm": 0.3200349807739258, + "learning_rate": 1.0656250000000002e-05, + "loss": 0.0162, + "step": 4590 + }, + { + "epoch": 47.043783956535634, + "grad_norm": 0.3281605541706085, + "learning_rate": 1.0635416666666667e-05, + "loss": 0.0158, + "step": 4600 + }, + { + "epoch": 47.043783956535634, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.256087303161621, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.744, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.413, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.052, + "step": 4600 + }, + { + "epoch": 47.14605305209332, + "grad_norm": 0.35677453875541687, + "learning_rate": 1.0614583333333335e-05, + "loss": 0.0153, + "step": 4610 + }, + { + "epoch": 47.24832214765101, + "grad_norm": 0.2933562099933624, + "learning_rate": 1.059375e-05, + "loss": 0.0151, + "step": 4620 + }, + { + "epoch": 47.35059124320869, + "grad_norm": 0.3412613272666931, + "learning_rate": 1.0572916666666668e-05, + "loss": 0.0154, + "step": 4630 + }, + { + "epoch": 47.45286033876638, + "grad_norm": 0.32954832911491394, + "learning_rate": 1.0552083333333333e-05, + "loss": 0.0154, + "step": 4640 + }, + { + "epoch": 47.555129434324066, + "grad_norm": 0.3282068073749542, + "learning_rate": 1.0531250000000002e-05, + "loss": 0.0152, + "step": 4650 + }, + { + "epoch": 47.65739852988175, + "grad_norm": 0.33588287234306335, + "learning_rate": 1.0510416666666667e-05, + "loss": 0.0156, + "step": 4660 + }, + { + "epoch": 47.75966762543944, + "grad_norm": 0.33600950241088867, + "learning_rate": 1.0489583333333335e-05, + "loss": 0.0151, + "step": 4670 + }, + { + "epoch": 47.86193672099712, + "grad_norm": 0.335001140832901, + "learning_rate": 1.046875e-05, + "loss": 0.0156, + "step": 4680 + }, + { + "epoch": 47.96420581655481, + "grad_norm": 0.33093252778053284, + "learning_rate": 1.0447916666666667e-05, + "loss": 0.0158, + "step": 4690 + }, + { + "epoch": 48.0664749121125, + "grad_norm": 0.32087039947509766, + "learning_rate": 1.0427083333333334e-05, + "loss": 0.0149, + "step": 4700 + }, + { + "epoch": 48.0664749121125, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.274885654449463, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8036, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.137, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.017, + "step": 4700 + }, + { + "epoch": 48.168744007670185, + "grad_norm": 0.31192752718925476, + "learning_rate": 1.040625e-05, + "loss": 0.0145, + "step": 4710 + }, + { + "epoch": 48.271013103227865, + "grad_norm": 0.2866131067276001, + "learning_rate": 1.0385416666666667e-05, + "loss": 0.0144, + "step": 4720 + }, + { + "epoch": 48.37328219878555, + "grad_norm": 0.34419262409210205, + "learning_rate": 1.0364583333333334e-05, + "loss": 0.0145, + "step": 4730 + }, + { + "epoch": 48.47555129434324, + "grad_norm": 0.33133459091186523, + "learning_rate": 1.034375e-05, + "loss": 0.0143, + "step": 4740 + }, + { + "epoch": 48.57782038990093, + "grad_norm": 0.3273964524269104, + "learning_rate": 1.0322916666666667e-05, + "loss": 0.0146, + "step": 4750 + }, + { + "epoch": 48.680089485458616, + "grad_norm": 0.3400149643421173, + "learning_rate": 1.0302083333333334e-05, + "loss": 0.0149, + "step": 4760 + }, + { + "epoch": 48.7823585810163, + "grad_norm": 0.3198685348033905, + "learning_rate": 1.0281250000000001e-05, + "loss": 0.0148, + "step": 4770 + }, + { + "epoch": 48.884627676573984, + "grad_norm": 0.3375503420829773, + "learning_rate": 1.0260416666666668e-05, + "loss": 0.0153, + "step": 4780 + }, + { + "epoch": 48.98689677213167, + "grad_norm": 0.3139088749885559, + "learning_rate": 1.0239583333333334e-05, + "loss": 0.0152, + "step": 4790 + }, + { + "epoch": 49.08916586768936, + "grad_norm": 0.32369062304496765, + "learning_rate": 1.0218750000000001e-05, + "loss": 0.0143, + "step": 4800 + }, + { + "epoch": 49.08916586768936, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2769861221313477, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7984, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.161, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.02, + "step": 4800 + }, + { + "epoch": 49.19143496324704, + "grad_norm": 0.3071221113204956, + "learning_rate": 1.0197916666666668e-05, + "loss": 0.0136, + "step": 4810 + }, + { + "epoch": 49.29370405880473, + "grad_norm": 0.31288015842437744, + "learning_rate": 1.0177083333333335e-05, + "loss": 0.0139, + "step": 4820 + }, + { + "epoch": 49.395973154362416, + "grad_norm": 0.31299805641174316, + "learning_rate": 1.0156250000000001e-05, + "loss": 0.0137, + "step": 4830 + }, + { + "epoch": 49.498242249920104, + "grad_norm": 0.3377828001976013, + "learning_rate": 1.0135416666666666e-05, + "loss": 0.0141, + "step": 4840 + }, + { + "epoch": 49.60051134547779, + "grad_norm": 0.31973740458488464, + "learning_rate": 1.0114583333333335e-05, + "loss": 0.0144, + "step": 4850 + }, + { + "epoch": 49.70278044103547, + "grad_norm": 0.3018786907196045, + "learning_rate": 1.009375e-05, + "loss": 0.0144, + "step": 4860 + }, + { + "epoch": 49.80504953659316, + "grad_norm": 0.3308105766773224, + "learning_rate": 1.0072916666666668e-05, + "loss": 0.0146, + "step": 4870 + }, + { + "epoch": 49.90731863215085, + "grad_norm": 0.3090561032295227, + "learning_rate": 1.0052083333333333e-05, + "loss": 0.0145, + "step": 4880 + }, + { + "epoch": 50.009587727708535, + "grad_norm": 0.33793970942497253, + "learning_rate": 1.0031250000000002e-05, + "loss": 0.0145, + "step": 4890 + }, + { + "epoch": 50.11185682326622, + "grad_norm": 0.3269507884979248, + "learning_rate": 1.0010416666666667e-05, + "loss": 0.0133, + "step": 4900 + }, + { + "epoch": 50.11185682326622, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2938148975372314, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7798, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.247, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.031, + "step": 4900 + }, + { + "epoch": 50.2141259188239, + "grad_norm": 0.30052992701530457, + "learning_rate": 9.989583333333333e-06, + "loss": 0.0136, + "step": 4910 + }, + { + "epoch": 50.31639501438159, + "grad_norm": 0.2977409362792969, + "learning_rate": 9.96875e-06, + "loss": 0.0134, + "step": 4920 + }, + { + "epoch": 50.41866410993928, + "grad_norm": 0.3022303283214569, + "learning_rate": 9.947916666666667e-06, + "loss": 0.0136, + "step": 4930 + }, + { + "epoch": 50.520933205496966, + "grad_norm": 0.2963425815105438, + "learning_rate": 9.927083333333334e-06, + "loss": 0.0137, + "step": 4940 + }, + { + "epoch": 50.62320230105465, + "grad_norm": 0.2862411141395569, + "learning_rate": 9.90625e-06, + "loss": 0.0134, + "step": 4950 + }, + { + "epoch": 50.725471396612335, + "grad_norm": 0.32539355754852295, + "learning_rate": 9.885416666666667e-06, + "loss": 0.0136, + "step": 4960 + }, + { + "epoch": 50.82774049217002, + "grad_norm": 0.30780264735221863, + "learning_rate": 9.864583333333334e-06, + "loss": 0.0138, + "step": 4970 + }, + { + "epoch": 50.93000958772771, + "grad_norm": 0.3183571994304657, + "learning_rate": 9.84375e-06, + "loss": 0.0136, + "step": 4980 + }, + { + "epoch": 51.0322786832854, + "grad_norm": 0.28671231865882874, + "learning_rate": 9.822916666666667e-06, + "loss": 0.0136, + "step": 4990 + }, + { + "epoch": 51.13454777884308, + "grad_norm": 0.2912745177745819, + "learning_rate": 9.802083333333334e-06, + "loss": 0.0126, + "step": 5000 + }, + { + "epoch": 51.13454777884308, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.312565326690674, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8065, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.124, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.015, + "step": 5000 + }, + { + "epoch": 51.236816874400766, + "grad_norm": 0.29751163721084595, + "learning_rate": 9.78125e-06, + "loss": 0.013, + "step": 5010 + }, + { + "epoch": 51.339085969958454, + "grad_norm": 0.30533960461616516, + "learning_rate": 9.760416666666667e-06, + "loss": 0.013, + "step": 5020 + }, + { + "epoch": 51.44135506551614, + "grad_norm": 0.3039548695087433, + "learning_rate": 9.739583333333334e-06, + "loss": 0.0129, + "step": 5030 + }, + { + "epoch": 51.54362416107382, + "grad_norm": 0.295386403799057, + "learning_rate": 9.71875e-06, + "loss": 0.0132, + "step": 5040 + }, + { + "epoch": 51.64589325663151, + "grad_norm": 0.289328932762146, + "learning_rate": 9.697916666666667e-06, + "loss": 0.0133, + "step": 5050 + }, + { + "epoch": 51.7481623521892, + "grad_norm": 0.3070317804813385, + "learning_rate": 9.677083333333334e-06, + "loss": 0.0131, + "step": 5060 + }, + { + "epoch": 51.850431447746885, + "grad_norm": 0.28922444581985474, + "learning_rate": 9.656250000000001e-06, + "loss": 0.0131, + "step": 5070 + }, + { + "epoch": 51.95270054330457, + "grad_norm": 0.3166629672050476, + "learning_rate": 9.635416666666668e-06, + "loss": 0.0132, + "step": 5080 + }, + { + "epoch": 52.054969638862254, + "grad_norm": 0.3101074993610382, + "learning_rate": 9.614583333333334e-06, + "loss": 0.0126, + "step": 5090 + }, + { + "epoch": 52.15723873441994, + "grad_norm": 0.318968266248703, + "learning_rate": 9.593750000000001e-06, + "loss": 0.0127, + "step": 5100 + }, + { + "epoch": 52.15723873441994, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3045780658721924, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8544, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.903, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.988, + "step": 5100 + }, + { + "epoch": 52.25950782997763, + "grad_norm": 0.30769291520118713, + "learning_rate": 9.572916666666668e-06, + "loss": 0.0122, + "step": 5110 + }, + { + "epoch": 52.36177692553532, + "grad_norm": 0.28409647941589355, + "learning_rate": 9.552083333333335e-06, + "loss": 0.0124, + "step": 5120 + }, + { + "epoch": 52.464046021093004, + "grad_norm": 0.3307039737701416, + "learning_rate": 9.531250000000001e-06, + "loss": 0.0128, + "step": 5130 + }, + { + "epoch": 52.566315116650685, + "grad_norm": 0.292473703622818, + "learning_rate": 9.510416666666668e-06, + "loss": 0.0125, + "step": 5140 + }, + { + "epoch": 52.66858421220837, + "grad_norm": 0.3042745888233185, + "learning_rate": 9.489583333333335e-06, + "loss": 0.0126, + "step": 5150 + }, + { + "epoch": 52.77085330776606, + "grad_norm": 0.29861128330230713, + "learning_rate": 9.468750000000001e-06, + "loss": 0.0128, + "step": 5160 + }, + { + "epoch": 52.87312240332375, + "grad_norm": 0.284404993057251, + "learning_rate": 9.447916666666668e-06, + "loss": 0.0129, + "step": 5170 + }, + { + "epoch": 52.97539149888143, + "grad_norm": 0.30096396803855896, + "learning_rate": 9.427083333333335e-06, + "loss": 0.0131, + "step": 5180 + }, + { + "epoch": 53.077660594439116, + "grad_norm": 0.3048815131187439, + "learning_rate": 9.406250000000002e-06, + "loss": 0.012, + "step": 5190 + }, + { + "epoch": 53.179929689996804, + "grad_norm": 0.2931344509124756, + "learning_rate": 9.385416666666668e-06, + "loss": 0.0118, + "step": 5200 + }, + { + "epoch": 53.179929689996804, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.321913480758667, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8669, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.845, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.981, + "step": 5200 + }, + { + "epoch": 53.28219878555449, + "grad_norm": 0.29786059260368347, + "learning_rate": 9.364583333333333e-06, + "loss": 0.0116, + "step": 5210 + }, + { + "epoch": 53.38446788111218, + "grad_norm": 0.2869616150856018, + "learning_rate": 9.34375e-06, + "loss": 0.0119, + "step": 5220 + }, + { + "epoch": 53.48673697666986, + "grad_norm": 0.3017300069332123, + "learning_rate": 9.322916666666667e-06, + "loss": 0.0121, + "step": 5230 + }, + { + "epoch": 53.58900607222755, + "grad_norm": 0.3326238691806793, + "learning_rate": 9.302083333333334e-06, + "loss": 0.0123, + "step": 5240 + }, + { + "epoch": 53.691275167785236, + "grad_norm": 0.30569109320640564, + "learning_rate": 9.28125e-06, + "loss": 0.0125, + "step": 5250 + }, + { + "epoch": 53.79354426334292, + "grad_norm": 0.3061717748641968, + "learning_rate": 9.260416666666667e-06, + "loss": 0.0123, + "step": 5260 + }, + { + "epoch": 53.895813358900604, + "grad_norm": 0.281955748796463, + "learning_rate": 9.239583333333334e-06, + "loss": 0.0121, + "step": 5270 + }, + { + "epoch": 53.99808245445829, + "grad_norm": 0.2920975983142853, + "learning_rate": 9.21875e-06, + "loss": 0.0124, + "step": 5280 + }, + { + "epoch": 54.10035155001598, + "grad_norm": 0.31395605206489563, + "learning_rate": 9.197916666666667e-06, + "loss": 0.0116, + "step": 5290 + }, + { + "epoch": 54.20262064557367, + "grad_norm": 0.3078666627407074, + "learning_rate": 9.177083333333334e-06, + "loss": 0.0117, + "step": 5300 + }, + { + "epoch": 54.20262064557367, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3260598182678223, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7898, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.201, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.025, + "step": 5300 + }, + { + "epoch": 54.304889741131355, + "grad_norm": 0.2747853994369507, + "learning_rate": 9.15625e-06, + "loss": 0.0113, + "step": 5310 + }, + { + "epoch": 54.407158836689035, + "grad_norm": 0.3068675398826599, + "learning_rate": 9.135416666666667e-06, + "loss": 0.0115, + "step": 5320 + }, + { + "epoch": 54.50942793224672, + "grad_norm": 0.29503345489501953, + "learning_rate": 9.114583333333334e-06, + "loss": 0.0116, + "step": 5330 + }, + { + "epoch": 54.61169702780441, + "grad_norm": 0.27636656165122986, + "learning_rate": 9.09375e-06, + "loss": 0.0114, + "step": 5340 + }, + { + "epoch": 54.7139661233621, + "grad_norm": 0.313203901052475, + "learning_rate": 9.072916666666668e-06, + "loss": 0.012, + "step": 5350 + }, + { + "epoch": 54.816235218919786, + "grad_norm": 0.30984997749328613, + "learning_rate": 9.052083333333334e-06, + "loss": 0.012, + "step": 5360 + }, + { + "epoch": 54.91850431447747, + "grad_norm": 0.27004456520080566, + "learning_rate": 9.031250000000001e-06, + "loss": 0.0118, + "step": 5370 + }, + { + "epoch": 55.020773410035154, + "grad_norm": 0.278777152299881, + "learning_rate": 9.010416666666668e-06, + "loss": 0.0117, + "step": 5380 + }, + { + "epoch": 55.12304250559284, + "grad_norm": 0.3070172965526581, + "learning_rate": 8.989583333333334e-06, + "loss": 0.011, + "step": 5390 + }, + { + "epoch": 55.22531160115053, + "grad_norm": 0.2756708562374115, + "learning_rate": 8.968750000000001e-06, + "loss": 0.011, + "step": 5400 + }, + { + "epoch": 55.22531160115053, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.33424711227417, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7849, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.224, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.028, + "step": 5400 + }, + { + "epoch": 55.32758069670821, + "grad_norm": 0.28067439794540405, + "learning_rate": 8.947916666666668e-06, + "loss": 0.011, + "step": 5410 + }, + { + "epoch": 55.4298497922659, + "grad_norm": 0.2702157199382782, + "learning_rate": 8.927083333333335e-06, + "loss": 0.0112, + "step": 5420 + }, + { + "epoch": 55.532118887823586, + "grad_norm": 0.30056026577949524, + "learning_rate": 8.906250000000001e-06, + "loss": 0.0115, + "step": 5430 + }, + { + "epoch": 55.634387983381274, + "grad_norm": 0.29118791222572327, + "learning_rate": 8.885416666666668e-06, + "loss": 0.0113, + "step": 5440 + }, + { + "epoch": 55.73665707893896, + "grad_norm": 0.3045700490474701, + "learning_rate": 8.864583333333335e-06, + "loss": 0.0115, + "step": 5450 + }, + { + "epoch": 55.83892617449664, + "grad_norm": 0.28140169382095337, + "learning_rate": 8.843750000000002e-06, + "loss": 0.0113, + "step": 5460 + }, + { + "epoch": 55.94119527005433, + "grad_norm": 0.2764737010002136, + "learning_rate": 8.822916666666668e-06, + "loss": 0.0118, + "step": 5470 + }, + { + "epoch": 56.04346436561202, + "grad_norm": 0.28735798597335815, + "learning_rate": 8.802083333333335e-06, + "loss": 0.0114, + "step": 5480 + }, + { + "epoch": 56.145733461169705, + "grad_norm": 0.28290146589279175, + "learning_rate": 8.781250000000002e-06, + "loss": 0.0107, + "step": 5490 + }, + { + "epoch": 56.248002556727386, + "grad_norm": 0.2917637526988983, + "learning_rate": 8.760416666666668e-06, + "loss": 0.0108, + "step": 5500 + }, + { + "epoch": 56.248002556727386, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3639163970947266, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7163, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.542, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.068, + "step": 5500 + }, + { + "epoch": 56.35027165228507, + "grad_norm": 0.31157800555229187, + "learning_rate": 8.739583333333333e-06, + "loss": 0.0109, + "step": 5510 + }, + { + "epoch": 56.45254074784276, + "grad_norm": 0.26494109630584717, + "learning_rate": 8.71875e-06, + "loss": 0.0108, + "step": 5520 + }, + { + "epoch": 56.55480984340045, + "grad_norm": 0.2839730381965637, + "learning_rate": 8.697916666666667e-06, + "loss": 0.011, + "step": 5530 + }, + { + "epoch": 56.657078938958136, + "grad_norm": 0.27168846130371094, + "learning_rate": 8.677083333333334e-06, + "loss": 0.0111, + "step": 5540 + }, + { + "epoch": 56.75934803451582, + "grad_norm": 0.2758902311325073, + "learning_rate": 8.65625e-06, + "loss": 0.0109, + "step": 5550 + }, + { + "epoch": 56.861617130073505, + "grad_norm": 0.29986515641212463, + "learning_rate": 8.635416666666667e-06, + "loss": 0.0113, + "step": 5560 + }, + { + "epoch": 56.96388622563119, + "grad_norm": 0.2895634174346924, + "learning_rate": 8.614583333333334e-06, + "loss": 0.0112, + "step": 5570 + }, + { + "epoch": 57.06615532118888, + "grad_norm": 0.26848530769348145, + "learning_rate": 8.59375e-06, + "loss": 0.0105, + "step": 5580 + }, + { + "epoch": 57.16842441674657, + "grad_norm": 0.2801918685436249, + "learning_rate": 8.572916666666667e-06, + "loss": 0.0106, + "step": 5590 + }, + { + "epoch": 57.27069351230425, + "grad_norm": 0.2772628366947174, + "learning_rate": 8.552083333333334e-06, + "loss": 0.0105, + "step": 5600 + }, + { + "epoch": 57.27069351230425, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3490209579467773, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7134, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.556, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.069, + "step": 5600 + }, + { + "epoch": 57.372962607861936, + "grad_norm": 0.26178139448165894, + "learning_rate": 8.53125e-06, + "loss": 0.0103, + "step": 5610 + }, + { + "epoch": 57.475231703419624, + "grad_norm": 0.27127423882484436, + "learning_rate": 8.510416666666667e-06, + "loss": 0.0105, + "step": 5620 + }, + { + "epoch": 57.57750079897731, + "grad_norm": 0.2728956639766693, + "learning_rate": 8.489583333333334e-06, + "loss": 0.0104, + "step": 5630 + }, + { + "epoch": 57.67976989453499, + "grad_norm": 0.26934826374053955, + "learning_rate": 8.468750000000001e-06, + "loss": 0.0107, + "step": 5640 + }, + { + "epoch": 57.78203899009268, + "grad_norm": 0.30065199732780457, + "learning_rate": 8.447916666666668e-06, + "loss": 0.0108, + "step": 5650 + }, + { + "epoch": 57.88430808565037, + "grad_norm": 0.29894405603408813, + "learning_rate": 8.427083333333334e-06, + "loss": 0.0104, + "step": 5660 + }, + { + "epoch": 57.986577181208055, + "grad_norm": 0.26421451568603516, + "learning_rate": 8.406250000000001e-06, + "loss": 0.0105, + "step": 5670 + }, + { + "epoch": 58.08884627676574, + "grad_norm": 0.27239277958869934, + "learning_rate": 8.385416666666668e-06, + "loss": 0.0099, + "step": 5680 + }, + { + "epoch": 58.191115372323424, + "grad_norm": 0.27293631434440613, + "learning_rate": 8.364583333333334e-06, + "loss": 0.01, + "step": 5690 + }, + { + "epoch": 58.29338446788111, + "grad_norm": 0.251788467168808, + "learning_rate": 8.343750000000001e-06, + "loss": 0.0103, + "step": 5700 + }, + { + "epoch": 58.29338446788111, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.360133647918701, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7658, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.312, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.039, + "step": 5700 + }, + { + "epoch": 58.3956535634388, + "grad_norm": 0.2742460072040558, + "learning_rate": 8.322916666666668e-06, + "loss": 0.01, + "step": 5710 + }, + { + "epoch": 58.49792265899649, + "grad_norm": 0.2849868834018707, + "learning_rate": 8.302083333333335e-06, + "loss": 0.0101, + "step": 5720 + }, + { + "epoch": 58.60019175455417, + "grad_norm": 0.2823048233985901, + "learning_rate": 8.281250000000001e-06, + "loss": 0.0101, + "step": 5730 + }, + { + "epoch": 58.702460850111855, + "grad_norm": 0.2816413342952728, + "learning_rate": 8.260416666666668e-06, + "loss": 0.0102, + "step": 5740 + }, + { + "epoch": 58.80472994566954, + "grad_norm": 0.3086257576942444, + "learning_rate": 8.239583333333335e-06, + "loss": 0.0102, + "step": 5750 + }, + { + "epoch": 58.90699904122723, + "grad_norm": 0.31176137924194336, + "learning_rate": 8.218750000000002e-06, + "loss": 0.0102, + "step": 5760 + }, + { + "epoch": 59.00926813678492, + "grad_norm": 0.30020809173583984, + "learning_rate": 8.197916666666668e-06, + "loss": 0.0103, + "step": 5770 + }, + { + "epoch": 59.1115372323426, + "grad_norm": 0.27167460322380066, + "learning_rate": 8.177083333333335e-06, + "loss": 0.0096, + "step": 5780 + }, + { + "epoch": 59.213806327900286, + "grad_norm": 0.29027628898620605, + "learning_rate": 8.156250000000002e-06, + "loss": 0.0099, + "step": 5790 + }, + { + "epoch": 59.316075423457974, + "grad_norm": 0.2502238154411316, + "learning_rate": 8.135416666666668e-06, + "loss": 0.0097, + "step": 5800 + }, + { + "epoch": 59.316075423457974, + "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3818840980529785, + "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7922, + "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.19, + "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.024, + "step": 5800 + } + ], + "logging_steps": 10, + "max_steps": 9700, + "num_input_tokens_seen": 0, + "num_train_epochs": 100, + "save_steps": 200, + "total_flos": 7.772644061346693e+18, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +}