diff --git "a/trainer_state.json" "b/trainer_state.json" deleted file mode 100644--- "a/trainer_state.json" +++ /dev/null @@ -1,4552 +0,0 @@ -{ - "best_metric": 2.3818840980529785, - "best_model_checkpoint": "/ssd1/geonmin.kim/shortened-llm/outputs/phi2_1.8b_alpaca_enzh_fullparam/checkpoint-5800", - "epoch": 59.316075423457974, - "eval_steps": 100, - "global_step": 5800, - "is_hyper_param_search": false, - "is_local_process_zero": true, - "is_world_process_zero": true, - "log_history": [ - { - "epoch": 0.010226909555768616, - "grad_norm": 3.242844581604004, - "learning_rate": 2.0000000000000002e-07, - "loss": 2.0435, - "step": 1 - }, - { - "epoch": 0.10226909555768617, - "grad_norm": 2.2696735858917236, - "learning_rate": 2.0000000000000003e-06, - "loss": 2.0397, - "step": 10 - }, - { - "epoch": 0.20453819111537233, - "grad_norm": 1.283516764640808, - "learning_rate": 4.000000000000001e-06, - "loss": 1.7851, - "step": 20 - }, - { - "epoch": 0.3068072866730585, - "grad_norm": 1.1238912343978882, - "learning_rate": 6e-06, - "loss": 1.6526, - "step": 30 - }, - { - "epoch": 0.40907638223074466, - "grad_norm": 1.0196110010147095, - "learning_rate": 8.000000000000001e-06, - "loss": 1.5391, - "step": 40 - }, - { - "epoch": 0.5113454777884308, - "grad_norm": 0.9775912761688232, - "learning_rate": 1e-05, - "loss": 1.4779, - "step": 50 - }, - { - "epoch": 0.613614573346117, - "grad_norm": 0.9060685038566589, - "learning_rate": 1.2e-05, - "loss": 1.4218, - "step": 60 - }, - { - "epoch": 0.7158836689038032, - "grad_norm": 0.9413829445838928, - "learning_rate": 1.4e-05, - "loss": 1.3978, - "step": 70 - }, - { - "epoch": 0.8181527644614893, - "grad_norm": 1.007612705230713, - "learning_rate": 1.6000000000000003e-05, - "loss": 1.3779, - "step": 80 - }, - { - "epoch": 0.9204218600191755, - "grad_norm": 0.9302048087120056, - "learning_rate": 1.8e-05, - "loss": 1.3387, - "step": 90 - }, - { - "epoch": 1.0226909555768615, - "grad_norm": 0.82750403881073, - "learning_rate": 2e-05, - "loss": 1.3056, - "step": 100 - }, - { - "epoch": 1.0226909555768615, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.201456069946289, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7448, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.41, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.051, - "step": 100 - }, - { - "epoch": 1.124960051134548, - "grad_norm": 0.8751218318939209, - "learning_rate": 1.9979166666666667e-05, - "loss": 1.2376, - "step": 110 - }, - { - "epoch": 1.2272291466922338, - "grad_norm": 0.8453890681266785, - "learning_rate": 1.9958333333333335e-05, - "loss": 1.2063, - "step": 120 - }, - { - "epoch": 1.3294982422499202, - "grad_norm": 0.8156936764717102, - "learning_rate": 1.99375e-05, - "loss": 1.2022, - "step": 130 - }, - { - "epoch": 1.4317673378076063, - "grad_norm": 0.8538714051246643, - "learning_rate": 1.991666666666667e-05, - "loss": 1.2052, - "step": 140 - }, - { - "epoch": 1.5340364333652925, - "grad_norm": 0.8687230944633484, - "learning_rate": 1.9895833333333334e-05, - "loss": 1.1913, - "step": 150 - }, - { - "epoch": 1.6363055289229784, - "grad_norm": 0.826032280921936, - "learning_rate": 1.9875000000000002e-05, - "loss": 1.1828, - "step": 160 - }, - { - "epoch": 1.7385746244806648, - "grad_norm": 0.8092572093009949, - "learning_rate": 1.9854166666666667e-05, - "loss": 1.1746, - "step": 170 - }, - { - "epoch": 1.840843720038351, - "grad_norm": 0.825042188167572, - "learning_rate": 1.9833333333333335e-05, - "loss": 1.169, - "step": 180 - }, - { - "epoch": 1.943112815596037, - "grad_norm": 0.8175085186958313, - "learning_rate": 1.98125e-05, - "loss": 1.166, - "step": 190 - }, - { - "epoch": 2.045381911153723, - "grad_norm": 0.8639950752258301, - "learning_rate": 1.979166666666667e-05, - "loss": 1.1158, - "step": 200 - }, - { - "epoch": 2.045381911153723, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1383644342422485, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7181, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.534, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.067, - "step": 200 - }, - { - "epoch": 2.1476510067114094, - "grad_norm": 0.7958649396896362, - "learning_rate": 1.9770833333333334e-05, - "loss": 1.0339, - "step": 210 - }, - { - "epoch": 2.249920102269096, - "grad_norm": 0.8010162711143494, - "learning_rate": 1.9750000000000002e-05, - "loss": 1.0426, - "step": 220 - }, - { - "epoch": 2.3521891978267817, - "grad_norm": 0.7885578274726868, - "learning_rate": 1.9729166666666667e-05, - "loss": 1.0432, - "step": 230 - }, - { - "epoch": 2.4544582933844676, - "grad_norm": 0.8269402980804443, - "learning_rate": 1.9708333333333336e-05, - "loss": 1.0379, - "step": 240 - }, - { - "epoch": 2.556727388942154, - "grad_norm": 0.8139219880104065, - "learning_rate": 1.96875e-05, - "loss": 1.0305, - "step": 250 - }, - { - "epoch": 2.6589964844998404, - "grad_norm": 0.8240389227867126, - "learning_rate": 1.9666666666666666e-05, - "loss": 1.032, - "step": 260 - }, - { - "epoch": 2.7612655800575263, - "grad_norm": 0.7902543544769287, - "learning_rate": 1.9645833333333334e-05, - "loss": 1.0203, - "step": 270 - }, - { - "epoch": 2.8635346756152127, - "grad_norm": 0.7832587361335754, - "learning_rate": 1.9625e-05, - "loss": 1.0266, - "step": 280 - }, - { - "epoch": 2.9658037711728986, - "grad_norm": 0.8064606189727783, - "learning_rate": 1.9604166666666668e-05, - "loss": 1.0292, - "step": 290 - }, - { - "epoch": 3.068072866730585, - "grad_norm": 0.8083460927009583, - "learning_rate": 1.9583333333333333e-05, - "loss": 0.9461, - "step": 300 - }, - { - "epoch": 3.068072866730585, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1164577007293701, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7799, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.247, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.031, - "step": 300 - }, - { - "epoch": 3.170341962288271, - "grad_norm": 0.7945894598960876, - "learning_rate": 1.95625e-05, - "loss": 0.9071, - "step": 310 - }, - { - "epoch": 3.2726110578459573, - "grad_norm": 0.8095683455467224, - "learning_rate": 1.9541666666666666e-05, - "loss": 0.9081, - "step": 320 - }, - { - "epoch": 3.3748801534036432, - "grad_norm": 0.779909074306488, - "learning_rate": 1.9520833333333335e-05, - "loss": 0.9171, - "step": 330 - }, - { - "epoch": 3.4771492489613296, - "grad_norm": 0.8173850774765015, - "learning_rate": 1.95e-05, - "loss": 0.9085, - "step": 340 - }, - { - "epoch": 3.5794183445190155, - "grad_norm": 0.8204581141471863, - "learning_rate": 1.9479166666666668e-05, - "loss": 0.9073, - "step": 350 - }, - { - "epoch": 3.681687440076702, - "grad_norm": 0.7985939383506775, - "learning_rate": 1.9458333333333333e-05, - "loss": 0.9181, - "step": 360 - }, - { - "epoch": 3.783956535634388, - "grad_norm": 0.8140767812728882, - "learning_rate": 1.94375e-05, - "loss": 0.9193, - "step": 370 - }, - { - "epoch": 3.886225631192074, - "grad_norm": 0.8102839589118958, - "learning_rate": 1.9416666666666667e-05, - "loss": 0.9157, - "step": 380 - }, - { - "epoch": 3.98849472674976, - "grad_norm": 0.7886559963226318, - "learning_rate": 1.9395833333333335e-05, - "loss": 0.9191, - "step": 390 - }, - { - "epoch": 4.090763822307446, - "grad_norm": 0.7990887761116028, - "learning_rate": 1.9375e-05, - "loss": 0.8203, - "step": 400 - }, - { - "epoch": 4.090763822307446, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1279124021530151, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7222, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.515, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.064, - "step": 400 - }, - { - "epoch": 4.193032917865133, - "grad_norm": 0.8134937286376953, - "learning_rate": 1.935416666666667e-05, - "loss": 0.81, - "step": 410 - }, - { - "epoch": 4.295302013422819, - "grad_norm": 0.8099086284637451, - "learning_rate": 1.9333333333333333e-05, - "loss": 0.8, - "step": 420 - }, - { - "epoch": 4.397571108980505, - "grad_norm": 0.7907654047012329, - "learning_rate": 1.9312500000000002e-05, - "loss": 0.8025, - "step": 430 - }, - { - "epoch": 4.499840204538192, - "grad_norm": 0.835167407989502, - "learning_rate": 1.9291666666666667e-05, - "loss": 0.804, - "step": 440 - }, - { - "epoch": 4.6021093000958775, - "grad_norm": 0.8259727358818054, - "learning_rate": 1.9270833333333335e-05, - "loss": 0.8069, - "step": 450 - }, - { - "epoch": 4.704378395653563, - "grad_norm": 0.8095329999923706, - "learning_rate": 1.925e-05, - "loss": 0.8099, - "step": 460 - }, - { - "epoch": 4.806647491211249, - "grad_norm": 0.8239871859550476, - "learning_rate": 1.922916666666667e-05, - "loss": 0.8178, - "step": 470 - }, - { - "epoch": 4.908916586768935, - "grad_norm": 0.8325999975204468, - "learning_rate": 1.9208333333333337e-05, - "loss": 0.8077, - "step": 480 - }, - { - "epoch": 5.011185682326622, - "grad_norm": 0.813887357711792, - "learning_rate": 1.9187500000000002e-05, - "loss": 0.8061, - "step": 490 - }, - { - "epoch": 5.113454777884308, - "grad_norm": 0.8288919925689697, - "learning_rate": 1.916666666666667e-05, - "loss": 0.7063, - "step": 500 - }, - { - "epoch": 5.113454777884308, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1451531648635864, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7807, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.243, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.03, - "step": 500 - }, - { - "epoch": 5.215723873441994, - "grad_norm": 0.8298630118370056, - "learning_rate": 1.9145833333333336e-05, - "loss": 0.7049, - "step": 510 - }, - { - "epoch": 5.317992968999681, - "grad_norm": 0.8246304988861084, - "learning_rate": 1.9125000000000004e-05, - "loss": 0.7104, - "step": 520 - }, - { - "epoch": 5.420262064557367, - "grad_norm": 0.7879628539085388, - "learning_rate": 1.910416666666667e-05, - "loss": 0.7147, - "step": 530 - }, - { - "epoch": 5.522531160115053, - "grad_norm": 0.8137240409851074, - "learning_rate": 1.9083333333333338e-05, - "loss": 0.7079, - "step": 540 - }, - { - "epoch": 5.624800255672739, - "grad_norm": 0.8201763033866882, - "learning_rate": 1.9062500000000003e-05, - "loss": 0.7162, - "step": 550 - }, - { - "epoch": 5.727069351230425, - "grad_norm": 0.8055077791213989, - "learning_rate": 1.9041666666666668e-05, - "loss": 0.7241, - "step": 560 - }, - { - "epoch": 5.829338446788111, - "grad_norm": 0.8504828810691833, - "learning_rate": 1.9020833333333336e-05, - "loss": 0.7241, - "step": 570 - }, - { - "epoch": 5.931607542345797, - "grad_norm": 0.8135042190551758, - "learning_rate": 1.9e-05, - "loss": 0.7276, - "step": 580 - }, - { - "epoch": 6.033876637903483, - "grad_norm": 0.907433271408081, - "learning_rate": 1.897916666666667e-05, - "loss": 0.6957, - "step": 590 - }, - { - "epoch": 6.13614573346117, - "grad_norm": 0.8348938822746277, - "learning_rate": 1.8958333333333334e-05, - "loss": 0.6172, - "step": 600 - }, - { - "epoch": 6.13614573346117, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.171920657157898, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7497, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.387, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.048, - "step": 600 - }, - { - "epoch": 6.238414829018856, - "grad_norm": 0.8306870460510254, - "learning_rate": 1.8937500000000003e-05, - "loss": 0.6255, - "step": 610 - }, - { - "epoch": 6.340683924576542, - "grad_norm": 0.8185003399848938, - "learning_rate": 1.8916666666666668e-05, - "loss": 0.6218, - "step": 620 - }, - { - "epoch": 6.442953020134228, - "grad_norm": 0.8396403789520264, - "learning_rate": 1.8895833333333336e-05, - "loss": 0.6282, - "step": 630 - }, - { - "epoch": 6.545222115691915, - "grad_norm": 0.8167343139648438, - "learning_rate": 1.8875e-05, - "loss": 0.6305, - "step": 640 - }, - { - "epoch": 6.6474912112496005, - "grad_norm": 0.8276931047439575, - "learning_rate": 1.885416666666667e-05, - "loss": 0.6378, - "step": 650 - }, - { - "epoch": 6.7497603068072864, - "grad_norm": 0.8405306339263916, - "learning_rate": 1.8833333333333335e-05, - "loss": 0.6375, - "step": 660 - }, - { - "epoch": 6.852029402364973, - "grad_norm": 0.8206018805503845, - "learning_rate": 1.8812500000000003e-05, - "loss": 0.6393, - "step": 670 - }, - { - "epoch": 6.954298497922659, - "grad_norm": 0.8008025884628296, - "learning_rate": 1.8791666666666668e-05, - "loss": 0.6448, - "step": 680 - }, - { - "epoch": 7.056567593480345, - "grad_norm": 0.8231362104415894, - "learning_rate": 1.8770833333333337e-05, - "loss": 0.5929, - "step": 690 - }, - { - "epoch": 7.158836689038031, - "grad_norm": 0.8469756841659546, - "learning_rate": 1.8750000000000002e-05, - "loss": 0.5392, - "step": 700 - }, - { - "epoch": 7.158836689038031, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.2256364822387695, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8506, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.92, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.99, - "step": 700 - }, - { - "epoch": 7.261105784595718, - "grad_norm": 0.8122310042381287, - "learning_rate": 1.8729166666666667e-05, - "loss": 0.54, - "step": 710 - }, - { - "epoch": 7.363374880153404, - "grad_norm": 0.8621034026145935, - "learning_rate": 1.8708333333333335e-05, - "loss": 0.5534, - "step": 720 - }, - { - "epoch": 7.46564397571109, - "grad_norm": 0.8529708981513977, - "learning_rate": 1.86875e-05, - "loss": 0.5516, - "step": 730 - }, - { - "epoch": 7.567913071268776, - "grad_norm": 0.8307532072067261, - "learning_rate": 1.866666666666667e-05, - "loss": 0.5531, - "step": 740 - }, - { - "epoch": 7.6701821668264625, - "grad_norm": 0.8403590321540833, - "learning_rate": 1.8645833333333334e-05, - "loss": 0.5644, - "step": 750 - }, - { - "epoch": 7.772451262384148, - "grad_norm": 0.8332897424697876, - "learning_rate": 1.8625000000000002e-05, - "loss": 0.5607, - "step": 760 - }, - { - "epoch": 7.874720357941834, - "grad_norm": 0.866201639175415, - "learning_rate": 1.8604166666666667e-05, - "loss": 0.5658, - "step": 770 - }, - { - "epoch": 7.97698945349952, - "grad_norm": 0.8362734913825989, - "learning_rate": 1.8583333333333336e-05, - "loss": 0.566, - "step": 780 - }, - { - "epoch": 8.079258549057206, - "grad_norm": 0.8327781558036804, - "learning_rate": 1.85625e-05, - "loss": 0.4963, - "step": 790 - }, - { - "epoch": 8.181527644614892, - "grad_norm": 0.827268123626709, - "learning_rate": 1.854166666666667e-05, - "loss": 0.4752, - "step": 800 - }, - { - "epoch": 8.181527644614892, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.2643427848815918, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.893, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.726, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.966, - "step": 800 - }, - { - "epoch": 8.28379674017258, - "grad_norm": 0.7902690172195435, - "learning_rate": 1.8520833333333334e-05, - "loss": 0.4843, - "step": 810 - }, - { - "epoch": 8.386065835730266, - "grad_norm": 0.8375966548919678, - "learning_rate": 1.8500000000000002e-05, - "loss": 0.4783, - "step": 820 - }, - { - "epoch": 8.488334931287952, - "grad_norm": 0.8248066306114197, - "learning_rate": 1.8479166666666667e-05, - "loss": 0.4785, - "step": 830 - }, - { - "epoch": 8.590604026845638, - "grad_norm": 0.8647124171257019, - "learning_rate": 1.8458333333333336e-05, - "loss": 0.4874, - "step": 840 - }, - { - "epoch": 8.692873122403324, - "grad_norm": 0.8471198678016663, - "learning_rate": 1.84375e-05, - "loss": 0.4884, - "step": 850 - }, - { - "epoch": 8.79514221796101, - "grad_norm": 0.8394653797149658, - "learning_rate": 1.8416666666666666e-05, - "loss": 0.497, - "step": 860 - }, - { - "epoch": 8.897411313518695, - "grad_norm": 0.8621006011962891, - "learning_rate": 1.8395833333333334e-05, - "loss": 0.4923, - "step": 870 - }, - { - "epoch": 8.999680409076383, - "grad_norm": 0.830193817615509, - "learning_rate": 1.8375e-05, - "loss": 0.4917, - "step": 880 - }, - { - "epoch": 9.101949504634069, - "grad_norm": 0.8418065309524536, - "learning_rate": 1.8354166666666668e-05, - "loss": 0.4104, - "step": 890 - }, - { - "epoch": 9.204218600191755, - "grad_norm": 0.8672240376472473, - "learning_rate": 1.8333333333333333e-05, - "loss": 0.4165, - "step": 900 - }, - { - "epoch": 9.204218600191755, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.2995867729187012, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8906, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.737, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.967, - "step": 900 - }, - { - "epoch": 9.306487695749441, - "grad_norm": 0.8118753433227539, - "learning_rate": 1.83125e-05, - "loss": 0.4181, - "step": 910 - }, - { - "epoch": 9.408756791307127, - "grad_norm": 0.8521497845649719, - "learning_rate": 1.8291666666666666e-05, - "loss": 0.4202, - "step": 920 - }, - { - "epoch": 9.511025886864813, - "grad_norm": 0.8396993279457092, - "learning_rate": 1.8270833333333335e-05, - "loss": 0.4247, - "step": 930 - }, - { - "epoch": 9.613294982422499, - "grad_norm": 0.8380371332168579, - "learning_rate": 1.825e-05, - "loss": 0.427, - "step": 940 - }, - { - "epoch": 9.715564077980185, - "grad_norm": 0.819240927696228, - "learning_rate": 1.8229166666666668e-05, - "loss": 0.4271, - "step": 950 - }, - { - "epoch": 9.817833173537872, - "grad_norm": 0.8467490077018738, - "learning_rate": 1.8208333333333333e-05, - "loss": 0.4288, - "step": 960 - }, - { - "epoch": 9.920102269095558, - "grad_norm": 0.8679558634757996, - "learning_rate": 1.81875e-05, - "loss": 0.4283, - "step": 970 - }, - { - "epoch": 10.022371364653244, - "grad_norm": 0.8336887359619141, - "learning_rate": 1.8166666666666667e-05, - "loss": 0.4244, - "step": 980 - }, - { - "epoch": 10.12464046021093, - "grad_norm": 0.8790706396102905, - "learning_rate": 1.8145833333333335e-05, - "loss": 0.3627, - "step": 990 - }, - { - "epoch": 10.226909555768616, - "grad_norm": 0.8260719180107117, - "learning_rate": 1.8125e-05, - "loss": 0.3622, - "step": 1000 - }, - { - "epoch": 10.226909555768616, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.3518736362457275, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.891, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.735, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.967, - "step": 1000 - }, - { - "epoch": 10.329178651326302, - "grad_norm": 0.8354102969169617, - "learning_rate": 1.810416666666667e-05, - "loss": 0.3654, - "step": 1010 - }, - { - "epoch": 10.431447746883988, - "grad_norm": 0.833742618560791, - "learning_rate": 1.8083333333333334e-05, - "loss": 0.3675, - "step": 1020 - }, - { - "epoch": 10.533716842441674, - "grad_norm": 0.814929723739624, - "learning_rate": 1.8062500000000002e-05, - "loss": 0.368, - "step": 1030 - }, - { - "epoch": 10.635985937999362, - "grad_norm": 0.8500260710716248, - "learning_rate": 1.8041666666666667e-05, - "loss": 0.3704, - "step": 1040 - }, - { - "epoch": 10.738255033557047, - "grad_norm": 0.8280666470527649, - "learning_rate": 1.8020833333333335e-05, - "loss": 0.3722, - "step": 1050 - }, - { - "epoch": 10.840524129114733, - "grad_norm": 0.871941089630127, - "learning_rate": 1.8e-05, - "loss": 0.3768, - "step": 1060 - }, - { - "epoch": 10.94279322467242, - "grad_norm": 0.8509662747383118, - "learning_rate": 1.797916666666667e-05, - "loss": 0.3829, - "step": 1070 - }, - { - "epoch": 11.045062320230105, - "grad_norm": 0.8621676564216614, - "learning_rate": 1.7958333333333334e-05, - "loss": 0.3512, - "step": 1080 - }, - { - "epoch": 11.147331415787791, - "grad_norm": 0.8415457010269165, - "learning_rate": 1.7937500000000002e-05, - "loss": 0.3092, - "step": 1090 - }, - { - "epoch": 11.249600511345477, - "grad_norm": 0.8087013363838196, - "learning_rate": 1.7916666666666667e-05, - "loss": 0.3158, - "step": 1100 - }, - { - "epoch": 11.249600511345477, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.4024713039398193, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8756, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.805, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.976, - "step": 1100 - }, - { - "epoch": 11.351869606903165, - "grad_norm": 0.8239210844039917, - "learning_rate": 1.7895833333333336e-05, - "loss": 0.3189, - "step": 1110 - }, - { - "epoch": 11.45413870246085, - "grad_norm": 0.8607499003410339, - "learning_rate": 1.7875e-05, - "loss": 0.3222, - "step": 1120 - }, - { - "epoch": 11.556407798018537, - "grad_norm": 0.8538540601730347, - "learning_rate": 1.785416666666667e-05, - "loss": 0.3207, - "step": 1130 - }, - { - "epoch": 11.658676893576223, - "grad_norm": 0.8388919234275818, - "learning_rate": 1.7833333333333334e-05, - "loss": 0.3308, - "step": 1140 - }, - { - "epoch": 11.760945989133909, - "grad_norm": 0.8096144795417786, - "learning_rate": 1.7812500000000003e-05, - "loss": 0.3256, - "step": 1150 - }, - { - "epoch": 11.863215084691594, - "grad_norm": 0.8199524879455566, - "learning_rate": 1.7791666666666668e-05, - "loss": 0.3272, - "step": 1160 - }, - { - "epoch": 11.96548418024928, - "grad_norm": 0.8503059148788452, - "learning_rate": 1.7770833333333336e-05, - "loss": 0.3313, - "step": 1170 - }, - { - "epoch": 12.067753275806966, - "grad_norm": 0.8376160860061646, - "learning_rate": 1.775e-05, - "loss": 0.2907, - "step": 1180 - }, - { - "epoch": 12.170022371364654, - "grad_norm": 5.976499557495117, - "learning_rate": 1.772916666666667e-05, - "loss": 0.2719, - "step": 1190 - }, - { - "epoch": 12.27229146692234, - "grad_norm": 0.8097366094589233, - "learning_rate": 1.7708333333333335e-05, - "loss": 0.2782, - "step": 1200 - }, - { - "epoch": 12.27229146692234, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.4481781721115112, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.9199, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.603, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.95, - "step": 1200 - }, - { - "epoch": 12.374560562480026, - "grad_norm": 0.7708520293235779, - "learning_rate": 1.7687500000000003e-05, - "loss": 0.2749, - "step": 1210 - }, - { - "epoch": 12.476829658037712, - "grad_norm": 0.8598915338516235, - "learning_rate": 1.7666666666666668e-05, - "loss": 0.2784, - "step": 1220 - }, - { - "epoch": 12.579098753595398, - "grad_norm": 0.8157161474227905, - "learning_rate": 1.7645833333333336e-05, - "loss": 0.2794, - "step": 1230 - }, - { - "epoch": 12.681367849153084, - "grad_norm": 0.8496010899543762, - "learning_rate": 1.7625e-05, - "loss": 0.2825, - "step": 1240 - }, - { - "epoch": 12.78363694471077, - "grad_norm": 0.815390944480896, - "learning_rate": 1.760416666666667e-05, - "loss": 0.2848, - "step": 1250 - }, - { - "epoch": 12.885906040268456, - "grad_norm": 0.8204165697097778, - "learning_rate": 1.7583333333333335e-05, - "loss": 0.2883, - "step": 1260 - }, - { - "epoch": 12.988175135826143, - "grad_norm": 0.8374896049499512, - "learning_rate": 1.7562500000000003e-05, - "loss": 0.2909, - "step": 1270 - }, - { - "epoch": 13.09044423138383, - "grad_norm": 0.7642439007759094, - "learning_rate": 1.754166666666667e-05, - "loss": 0.2406, - "step": 1280 - }, - { - "epoch": 13.192713326941515, - "grad_norm": 0.7860681414604187, - "learning_rate": 1.7520833333333337e-05, - "loss": 0.2358, - "step": 1290 - }, - { - "epoch": 13.294982422499201, - "grad_norm": 0.8119321465492249, - "learning_rate": 1.7500000000000002e-05, - "loss": 0.2357, - "step": 1300 - }, - { - "epoch": 13.294982422499201, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.4959287643432617, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8771, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.799, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.975, - "step": 1300 - }, - { - "epoch": 13.397251518056887, - "grad_norm": 0.7898840308189392, - "learning_rate": 1.7479166666666667e-05, - "loss": 0.2409, - "step": 1310 - }, - { - "epoch": 13.499520613614573, - "grad_norm": 0.8298600912094116, - "learning_rate": 1.7458333333333335e-05, - "loss": 0.2399, - "step": 1320 - }, - { - "epoch": 13.601789709172259, - "grad_norm": 0.7994723916053772, - "learning_rate": 1.74375e-05, - "loss": 0.2468, - "step": 1330 - }, - { - "epoch": 13.704058804729947, - "grad_norm": 0.822475790977478, - "learning_rate": 1.741666666666667e-05, - "loss": 0.2481, - "step": 1340 - }, - { - "epoch": 13.806327900287632, - "grad_norm": 0.8012453317642212, - "learning_rate": 1.7395833333333334e-05, - "loss": 0.2473, - "step": 1350 - }, - { - "epoch": 13.908596995845318, - "grad_norm": 0.8046063780784607, - "learning_rate": 1.7375000000000002e-05, - "loss": 0.2532, - "step": 1360 - }, - { - "epoch": 14.010866091403004, - "grad_norm": 0.7180681228637695, - "learning_rate": 1.7354166666666667e-05, - "loss": 0.2451, - "step": 1370 - }, - { - "epoch": 14.11313518696069, - "grad_norm": 0.7648767828941345, - "learning_rate": 1.7333333333333336e-05, - "loss": 0.2026, - "step": 1380 - }, - { - "epoch": 14.215404282518376, - "grad_norm": 0.782311201095581, - "learning_rate": 1.73125e-05, - "loss": 0.2062, - "step": 1390 - }, - { - "epoch": 14.317673378076062, - "grad_norm": 0.7766838669776917, - "learning_rate": 1.729166666666667e-05, - "loss": 0.2079, - "step": 1400 - }, - { - "epoch": 14.317673378076062, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.535245418548584, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8983, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.701, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.963, - "step": 1400 - }, - { - "epoch": 14.419942473633748, - "grad_norm": 0.7595117688179016, - "learning_rate": 1.7270833333333334e-05, - "loss": 0.2097, - "step": 1410 - }, - { - "epoch": 14.522211569191436, - "grad_norm": 0.7640486359596252, - "learning_rate": 1.7250000000000003e-05, - "loss": 0.2107, - "step": 1420 - }, - { - "epoch": 14.624480664749122, - "grad_norm": 0.8100217580795288, - "learning_rate": 1.7229166666666668e-05, - "loss": 0.2132, - "step": 1430 - }, - { - "epoch": 14.726749760306808, - "grad_norm": 0.7824357748031616, - "learning_rate": 1.7208333333333336e-05, - "loss": 0.2124, - "step": 1440 - }, - { - "epoch": 14.829018855864494, - "grad_norm": 0.8215783834457397, - "learning_rate": 1.71875e-05, - "loss": 0.2172, - "step": 1450 - }, - { - "epoch": 14.93128795142218, - "grad_norm": 0.791244626045227, - "learning_rate": 1.7166666666666666e-05, - "loss": 0.2183, - "step": 1460 - }, - { - "epoch": 15.033557046979865, - "grad_norm": 0.7542420625686646, - "learning_rate": 1.7145833333333334e-05, - "loss": 0.2083, - "step": 1470 - }, - { - "epoch": 15.135826142537551, - "grad_norm": 0.7965226173400879, - "learning_rate": 1.7125e-05, - "loss": 0.1773, - "step": 1480 - }, - { - "epoch": 15.238095238095237, - "grad_norm": 0.764574408531189, - "learning_rate": 1.7104166666666668e-05, - "loss": 0.1775, - "step": 1490 - }, - { - "epoch": 15.340364333652925, - "grad_norm": 0.768020749092102, - "learning_rate": 1.7083333333333333e-05, - "loss": 0.181, - "step": 1500 - }, - { - "epoch": 15.340364333652925, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.57305908203125, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7396, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.434, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.054, - "step": 1500 - }, - { - "epoch": 15.44263342921061, - "grad_norm": 0.7567213773727417, - "learning_rate": 1.70625e-05, - "loss": 0.181, - "step": 1510 - }, - { - "epoch": 15.544902524768297, - "grad_norm": 0.7426446080207825, - "learning_rate": 1.7041666666666666e-05, - "loss": 0.1824, - "step": 1520 - }, - { - "epoch": 15.647171620325983, - "grad_norm": 0.750170111656189, - "learning_rate": 1.7020833333333335e-05, - "loss": 0.1855, - "step": 1530 - }, - { - "epoch": 15.749440715883669, - "grad_norm": 0.7680428624153137, - "learning_rate": 1.7e-05, - "loss": 0.1883, - "step": 1540 - }, - { - "epoch": 15.851709811441355, - "grad_norm": 0.8081603646278381, - "learning_rate": 1.6979166666666668e-05, - "loss": 0.1892, - "step": 1550 - }, - { - "epoch": 15.95397890699904, - "grad_norm": 0.7882938385009766, - "learning_rate": 1.6958333333333333e-05, - "loss": 0.1912, - "step": 1560 - }, - { - "epoch": 16.05624800255673, - "grad_norm": 0.7245915532112122, - "learning_rate": 1.6937500000000002e-05, - "loss": 0.1717, - "step": 1570 - }, - { - "epoch": 16.158517098114412, - "grad_norm": 0.7572883367538452, - "learning_rate": 1.6916666666666667e-05, - "loss": 0.1577, - "step": 1580 - }, - { - "epoch": 16.2607861936721, - "grad_norm": 0.7576362490653992, - "learning_rate": 1.6895833333333335e-05, - "loss": 0.157, - "step": 1590 - }, - { - "epoch": 16.363055289229784, - "grad_norm": 0.7182960510253906, - "learning_rate": 1.6875e-05, - "loss": 0.1555, - "step": 1600 - }, - { - "epoch": 16.363055289229784, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.6099534034729004, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8255, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.036, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.005, - "step": 1600 - }, - { - "epoch": 16.465324384787472, - "grad_norm": 0.7401660084724426, - "learning_rate": 1.685416666666667e-05, - "loss": 0.1573, - "step": 1610 - }, - { - "epoch": 16.56759348034516, - "grad_norm": 0.7596891522407532, - "learning_rate": 1.6833333333333334e-05, - "loss": 0.1589, - "step": 1620 - }, - { - "epoch": 16.669862575902844, - "grad_norm": 0.767320990562439, - "learning_rate": 1.6812500000000002e-05, - "loss": 0.1613, - "step": 1630 - }, - { - "epoch": 16.77213167146053, - "grad_norm": 0.7579568028450012, - "learning_rate": 1.6791666666666667e-05, - "loss": 0.1624, - "step": 1640 - }, - { - "epoch": 16.874400767018216, - "grad_norm": 0.7488529682159424, - "learning_rate": 1.6770833333333336e-05, - "loss": 0.1646, - "step": 1650 - }, - { - "epoch": 16.976669862575903, - "grad_norm": 0.7528676390647888, - "learning_rate": 1.675e-05, - "loss": 0.1686, - "step": 1660 - }, - { - "epoch": 17.078938958133588, - "grad_norm": 0.6930709481239319, - "learning_rate": 1.672916666666667e-05, - "loss": 0.1429, - "step": 1670 - }, - { - "epoch": 17.181208053691275, - "grad_norm": 0.7119250297546387, - "learning_rate": 1.6708333333333334e-05, - "loss": 0.1355, - "step": 1680 - }, - { - "epoch": 17.283477149248963, - "grad_norm": 0.6774548292160034, - "learning_rate": 1.6687500000000002e-05, - "loss": 0.1348, - "step": 1690 - }, - { - "epoch": 17.385746244806647, - "grad_norm": 0.7192063927650452, - "learning_rate": 1.6666666666666667e-05, - "loss": 0.1368, - "step": 1700 - }, - { - "epoch": 17.385746244806647, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.6659198999404907, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8832, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.771, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.971, - "step": 1700 - }, - { - "epoch": 17.488015340364335, - "grad_norm": 0.6981900930404663, - "learning_rate": 1.6645833333333336e-05, - "loss": 0.1379, - "step": 1710 - }, - { - "epoch": 17.59028443592202, - "grad_norm": 0.7018482685089111, - "learning_rate": 1.6625e-05, - "loss": 0.1413, - "step": 1720 - }, - { - "epoch": 17.692553531479707, - "grad_norm": 0.7282826900482178, - "learning_rate": 1.660416666666667e-05, - "loss": 0.1437, - "step": 1730 - }, - { - "epoch": 17.79482262703739, - "grad_norm": 0.751104474067688, - "learning_rate": 1.6583333333333334e-05, - "loss": 0.142, - "step": 1740 - }, - { - "epoch": 17.89709172259508, - "grad_norm": 0.7333133816719055, - "learning_rate": 1.6562500000000003e-05, - "loss": 0.1452, - "step": 1750 - }, - { - "epoch": 17.999360818152766, - "grad_norm": 0.7490417957305908, - "learning_rate": 1.6541666666666668e-05, - "loss": 0.1445, - "step": 1760 - }, - { - "epoch": 18.10162991371045, - "grad_norm": 0.7038053274154663, - "learning_rate": 1.6520833333333336e-05, - "loss": 0.1196, - "step": 1770 - }, - { - "epoch": 18.203899009268138, - "grad_norm": 0.6416111588478088, - "learning_rate": 1.65e-05, - "loss": 0.1189, - "step": 1780 - }, - { - "epoch": 18.306168104825822, - "grad_norm": 0.6799057126045227, - "learning_rate": 1.647916666666667e-05, - "loss": 0.1187, - "step": 1790 - }, - { - "epoch": 18.40843720038351, - "grad_norm": 0.70688396692276, - "learning_rate": 1.6458333333333335e-05, - "loss": 0.1195, - "step": 1800 - }, - { - "epoch": 18.40843720038351, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.701189398765564, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8449, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.947, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.993, - "step": 1800 - }, - { - "epoch": 18.510706295941194, - "grad_norm": 0.6970006823539734, - "learning_rate": 1.6437500000000003e-05, - "loss": 0.1223, - "step": 1810 - }, - { - "epoch": 18.612975391498882, - "grad_norm": 0.7127917408943176, - "learning_rate": 1.6416666666666668e-05, - "loss": 0.1228, - "step": 1820 - }, - { - "epoch": 18.715244487056566, - "grad_norm": 0.6875537633895874, - "learning_rate": 1.6395833333333337e-05, - "loss": 0.1234, - "step": 1830 - }, - { - "epoch": 18.817513582614254, - "grad_norm": 0.7249884009361267, - "learning_rate": 1.6375e-05, - "loss": 0.126, - "step": 1840 - }, - { - "epoch": 18.91978267817194, - "grad_norm": 0.7007323503494263, - "learning_rate": 1.635416666666667e-05, - "loss": 0.1243, - "step": 1850 - }, - { - "epoch": 19.022051773729626, - "grad_norm": 0.6626154184341431, - "learning_rate": 1.6333333333333335e-05, - "loss": 0.1233, - "step": 1860 - }, - { - "epoch": 19.124320869287313, - "grad_norm": 0.6599249243736267, - "learning_rate": 1.6312500000000003e-05, - "loss": 0.1034, - "step": 1870 - }, - { - "epoch": 19.226589964844997, - "grad_norm": 0.6787338256835938, - "learning_rate": 1.629166666666667e-05, - "loss": 0.1031, - "step": 1880 - }, - { - "epoch": 19.328859060402685, - "grad_norm": 0.6463894248008728, - "learning_rate": 1.6270833333333337e-05, - "loss": 0.106, - "step": 1890 - }, - { - "epoch": 19.43112815596037, - "grad_norm": 0.6649991869926453, - "learning_rate": 1.6250000000000002e-05, - "loss": 0.1063, - "step": 1900 - }, - { - "epoch": 19.43112815596037, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.7399890422821045, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8883, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.747, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.968, - "step": 1900 - }, - { - "epoch": 19.533397251518057, - "grad_norm": 0.6542336940765381, - "learning_rate": 1.6229166666666667e-05, - "loss": 0.1059, - "step": 1910 - }, - { - "epoch": 19.635666347075745, - "grad_norm": 0.6741412281990051, - "learning_rate": 1.6208333333333335e-05, - "loss": 0.1088, - "step": 1920 - }, - { - "epoch": 19.73793544263343, - "grad_norm": 0.6898075938224792, - "learning_rate": 1.61875e-05, - "loss": 0.1087, - "step": 1930 - }, - { - "epoch": 19.840204538191117, - "grad_norm": 0.6611754298210144, - "learning_rate": 1.616666666666667e-05, - "loss": 0.1098, - "step": 1940 - }, - { - "epoch": 19.9424736337488, - "grad_norm": 0.6986993551254272, - "learning_rate": 1.6145833333333334e-05, - "loss": 0.1114, - "step": 1950 - }, - { - "epoch": 20.04474272930649, - "grad_norm": 0.6218557953834534, - "learning_rate": 1.6125000000000002e-05, - "loss": 0.1028, - "step": 1960 - }, - { - "epoch": 20.147011824864173, - "grad_norm": 0.6389386653900146, - "learning_rate": 1.6104166666666667e-05, - "loss": 0.0922, - "step": 1970 - }, - { - "epoch": 20.24928092042186, - "grad_norm": 0.6259350776672363, - "learning_rate": 1.6083333333333336e-05, - "loss": 0.0924, - "step": 1980 - }, - { - "epoch": 20.351550015979548, - "grad_norm": 0.6429017782211304, - "learning_rate": 1.60625e-05, - "loss": 0.0935, - "step": 1990 - }, - { - "epoch": 20.453819111537232, - "grad_norm": 0.6409590244293213, - "learning_rate": 1.604166666666667e-05, - "loss": 0.0925, - "step": 2000 - }, - { - "epoch": 20.453819111537232, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.770869493484497, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8821, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.776, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.972, - "step": 2000 - }, - { - "epoch": 20.55608820709492, - "grad_norm": 0.6365486979484558, - "learning_rate": 1.6020833333333334e-05, - "loss": 0.0945, - "step": 2010 - }, - { - "epoch": 20.658357302652604, - "grad_norm": 0.6637431979179382, - "learning_rate": 1.6000000000000003e-05, - "loss": 0.0963, - "step": 2020 - }, - { - "epoch": 20.76062639821029, - "grad_norm": 0.6256012916564941, - "learning_rate": 1.5979166666666668e-05, - "loss": 0.0981, - "step": 2030 - }, - { - "epoch": 20.862895493767976, - "grad_norm": 0.6379542946815491, - "learning_rate": 1.5958333333333336e-05, - "loss": 0.0976, - "step": 2040 - }, - { - "epoch": 20.965164589325664, - "grad_norm": 0.6680212020874023, - "learning_rate": 1.59375e-05, - "loss": 0.0982, - "step": 2050 - }, - { - "epoch": 21.067433684883348, - "grad_norm": 0.6050080060958862, - "learning_rate": 1.5916666666666666e-05, - "loss": 0.0876, - "step": 2060 - }, - { - "epoch": 21.169702780441035, - "grad_norm": 0.6315256953239441, - "learning_rate": 1.5895833333333335e-05, - "loss": 0.0824, - "step": 2070 - }, - { - "epoch": 21.271971875998723, - "grad_norm": 0.6160369515419006, - "learning_rate": 1.5875e-05, - "loss": 0.0817, - "step": 2080 - }, - { - "epoch": 21.374240971556407, - "grad_norm": 0.5967450737953186, - "learning_rate": 1.5854166666666668e-05, - "loss": 0.0828, - "step": 2090 - }, - { - "epoch": 21.476510067114095, - "grad_norm": 0.6033092737197876, - "learning_rate": 1.5833333333333333e-05, - "loss": 0.0832, - "step": 2100 - }, - { - "epoch": 21.476510067114095, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.8033993244171143, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8632, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.862, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.983, - "step": 2100 - }, - { - "epoch": 21.57877916267178, - "grad_norm": 0.6197848916053772, - "learning_rate": 1.58125e-05, - "loss": 0.0852, - "step": 2110 - }, - { - "epoch": 21.681048258229467, - "grad_norm": 0.6332718133926392, - "learning_rate": 1.5791666666666667e-05, - "loss": 0.0853, - "step": 2120 - }, - { - "epoch": 21.78331735378715, - "grad_norm": 0.6289674043655396, - "learning_rate": 1.5770833333333335e-05, - "loss": 0.0861, - "step": 2130 - }, - { - "epoch": 21.88558644934484, - "grad_norm": 0.6466374397277832, - "learning_rate": 1.575e-05, - "loss": 0.0873, - "step": 2140 - }, - { - "epoch": 21.987855544902526, - "grad_norm": 2.8282721042633057, - "learning_rate": 1.5733333333333334e-05, - "loss": 0.0888, - "step": 2150 - }, - { - "epoch": 22.09012464046021, - "grad_norm": 0.5769690275192261, - "learning_rate": 1.5712500000000002e-05, - "loss": 0.0748, - "step": 2160 - }, - { - "epoch": 22.192393736017898, - "grad_norm": 0.5819457173347473, - "learning_rate": 1.5691666666666667e-05, - "loss": 0.0737, - "step": 2170 - }, - { - "epoch": 22.294662831575582, - "grad_norm": 0.6134530305862427, - "learning_rate": 1.5670833333333336e-05, - "loss": 0.0739, - "step": 2180 - }, - { - "epoch": 22.39693192713327, - "grad_norm": 0.6075708866119385, - "learning_rate": 1.565e-05, - "loss": 0.0742, - "step": 2190 - }, - { - "epoch": 22.499201022690954, - "grad_norm": 0.5924075245857239, - "learning_rate": 1.562916666666667e-05, - "loss": 0.0751, - "step": 2200 - }, - { - "epoch": 22.499201022690954, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.829688549041748, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8883, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.747, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.968, - "step": 2200 - }, - { - "epoch": 22.601470118248642, - "grad_norm": 0.6079824566841125, - "learning_rate": 1.5608333333333334e-05, - "loss": 0.0762, - "step": 2210 - }, - { - "epoch": 22.70373921380633, - "grad_norm": 0.6125743389129639, - "learning_rate": 1.5587500000000003e-05, - "loss": 0.0764, - "step": 2220 - }, - { - "epoch": 22.806008309364014, - "grad_norm": 0.5798956751823425, - "learning_rate": 1.5566666666666668e-05, - "loss": 0.0764, - "step": 2230 - }, - { - "epoch": 22.9082774049217, - "grad_norm": 0.6115660667419434, - "learning_rate": 1.5545833333333336e-05, - "loss": 0.0782, - "step": 2240 - }, - { - "epoch": 23.010546500479386, - "grad_norm": 0.5898561477661133, - "learning_rate": 1.5525e-05, - "loss": 0.0785, - "step": 2250 - }, - { - "epoch": 23.112815596037073, - "grad_norm": 0.5677723288536072, - "learning_rate": 1.550416666666667e-05, - "loss": 0.066, - "step": 2260 - }, - { - "epoch": 23.215084691594758, - "grad_norm": 0.5659050345420837, - "learning_rate": 1.5483333333333335e-05, - "loss": 0.0659, - "step": 2270 - }, - { - "epoch": 23.317353787152445, - "grad_norm": 0.5879548192024231, - "learning_rate": 1.54625e-05, - "loss": 0.0671, - "step": 2280 - }, - { - "epoch": 23.41962288271013, - "grad_norm": 0.5900695323944092, - "learning_rate": 1.5441666666666668e-05, - "loss": 0.0668, - "step": 2290 - }, - { - "epoch": 23.521891978267817, - "grad_norm": 0.5663672685623169, - "learning_rate": 1.5420833333333333e-05, - "loss": 0.0676, - "step": 2300 - }, - { - "epoch": 23.521891978267817, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.8641977310180664, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8705, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.829, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.979, - "step": 2300 - }, - { - "epoch": 23.624161073825505, - "grad_norm": 0.6123341917991638, - "learning_rate": 1.54e-05, - "loss": 0.0686, - "step": 2310 - }, - { - "epoch": 23.72643016938319, - "grad_norm": 0.5857561826705933, - "learning_rate": 1.5379166666666667e-05, - "loss": 0.0688, - "step": 2320 - }, - { - "epoch": 23.828699264940877, - "grad_norm": 0.6261973977088928, - "learning_rate": 1.5358333333333335e-05, - "loss": 0.0694, - "step": 2330 - }, - { - "epoch": 23.93096836049856, - "grad_norm": 0.5833300948143005, - "learning_rate": 1.53375e-05, - "loss": 0.0706, - "step": 2340 - }, - { - "epoch": 24.03323745605625, - "grad_norm": 0.5474048256874084, - "learning_rate": 1.531666666666667e-05, - "loss": 0.0676, - "step": 2350 - }, - { - "epoch": 24.135506551613933, - "grad_norm": 0.5484219193458557, - "learning_rate": 1.5295833333333334e-05, - "loss": 0.06, - "step": 2360 - }, - { - "epoch": 24.23777564717162, - "grad_norm": 0.5514199137687683, - "learning_rate": 1.5275000000000002e-05, - "loss": 0.0608, - "step": 2370 - }, - { - "epoch": 24.340044742729308, - "grad_norm": 0.5593263506889343, - "learning_rate": 1.5254166666666667e-05, - "loss": 0.0607, - "step": 2380 - }, - { - "epoch": 24.442313838286992, - "grad_norm": 0.5621811151504517, - "learning_rate": 1.5233333333333335e-05, - "loss": 0.0612, - "step": 2390 - }, - { - "epoch": 24.54458293384468, - "grad_norm": 0.5351541042327881, - "learning_rate": 1.52125e-05, - "loss": 0.0612, - "step": 2400 - }, - { - "epoch": 24.54458293384468, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.8911913633346558, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8888, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.745, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.968, - "step": 2400 - }, - { - "epoch": 24.646852029402364, - "grad_norm": 0.6047305464744568, - "learning_rate": 1.5191666666666669e-05, - "loss": 0.0628, - "step": 2410 - }, - { - "epoch": 24.749121124960052, - "grad_norm": 0.5636226534843445, - "learning_rate": 1.5170833333333334e-05, - "loss": 0.0624, - "step": 2420 - }, - { - "epoch": 24.851390220517736, - "grad_norm": 0.5656375288963318, - "learning_rate": 1.515e-05, - "loss": 0.0635, - "step": 2430 - }, - { - "epoch": 24.953659316075424, - "grad_norm": 0.5728236436843872, - "learning_rate": 1.5129166666666667e-05, - "loss": 0.0642, - "step": 2440 - }, - { - "epoch": 25.05592841163311, - "grad_norm": 0.5383201837539673, - "learning_rate": 1.5108333333333334e-05, - "loss": 0.0585, - "step": 2450 - }, - { - "epoch": 25.158197507190796, - "grad_norm": 0.5213799476623535, - "learning_rate": 1.50875e-05, - "loss": 0.0539, - "step": 2460 - }, - { - "epoch": 25.260466602748483, - "grad_norm": 0.5278561115264893, - "learning_rate": 1.5066666666666668e-05, - "loss": 0.0557, - "step": 2470 - }, - { - "epoch": 25.362735698306167, - "grad_norm": 0.5517110824584961, - "learning_rate": 1.5045833333333334e-05, - "loss": 0.054, - "step": 2480 - }, - { - "epoch": 25.465004793863855, - "grad_norm": 0.5266678333282471, - "learning_rate": 1.5025000000000001e-05, - "loss": 0.055, - "step": 2490 - }, - { - "epoch": 25.56727388942154, - "grad_norm": 0.5280548334121704, - "learning_rate": 1.5004166666666668e-05, - "loss": 0.056, - "step": 2500 - }, - { - "epoch": 25.56727388942154, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.9079244136810303, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8724, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.82, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.978, - "step": 2500 - }, - { - "epoch": 25.669542984979227, - "grad_norm": 0.543380081653595, - "learning_rate": 1.4983333333333334e-05, - "loss": 0.0567, - "step": 2510 - }, - { - "epoch": 25.77181208053691, - "grad_norm": 0.5638805031776428, - "learning_rate": 1.4962500000000003e-05, - "loss": 0.0576, - "step": 2520 - }, - { - "epoch": 25.8740811760946, - "grad_norm": 0.5791529417037964, - "learning_rate": 1.4941666666666668e-05, - "loss": 0.0585, - "step": 2530 - }, - { - "epoch": 25.976350271652286, - "grad_norm": 0.5550780892372131, - "learning_rate": 1.4920833333333336e-05, - "loss": 0.0583, - "step": 2540 - }, - { - "epoch": 26.07861936720997, - "grad_norm": 0.5031822323799133, - "learning_rate": 1.4900000000000001e-05, - "loss": 0.0521, - "step": 2550 - }, - { - "epoch": 26.18088846276766, - "grad_norm": 0.5042924880981445, - "learning_rate": 1.487916666666667e-05, - "loss": 0.0495, - "step": 2560 - }, - { - "epoch": 26.283157558325343, - "grad_norm": 0.5483749508857727, - "learning_rate": 1.4858333333333335e-05, - "loss": 0.0504, - "step": 2570 - }, - { - "epoch": 26.38542665388303, - "grad_norm": 0.508752703666687, - "learning_rate": 1.48375e-05, - "loss": 0.0509, - "step": 2580 - }, - { - "epoch": 26.487695749440714, - "grad_norm": 0.4737156331539154, - "learning_rate": 1.4816666666666668e-05, - "loss": 0.0507, - "step": 2590 - }, - { - "epoch": 26.589964844998402, - "grad_norm": 0.5402019023895264, - "learning_rate": 1.4795833333333333e-05, - "loss": 0.0513, - "step": 2600 - }, - { - "epoch": 26.589964844998402, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.9507412910461426, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8703, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.83, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.979, - "step": 2600 - }, - { - "epoch": 26.69223394055609, - "grad_norm": 0.5165488123893738, - "learning_rate": 1.4775000000000002e-05, - "loss": 0.0519, - "step": 2610 - }, - { - "epoch": 26.794503036113774, - "grad_norm": 0.5218144059181213, - "learning_rate": 1.4754166666666667e-05, - "loss": 0.0537, - "step": 2620 - }, - { - "epoch": 26.89677213167146, - "grad_norm": 0.5503118634223938, - "learning_rate": 1.4733333333333335e-05, - "loss": 0.053, - "step": 2630 - }, - { - "epoch": 26.999041227229146, - "grad_norm": 0.5653933882713318, - "learning_rate": 1.47125e-05, - "loss": 0.0533, - "step": 2640 - }, - { - "epoch": 27.101310322786833, - "grad_norm": 0.5077876448631287, - "learning_rate": 1.4691666666666669e-05, - "loss": 0.0473, - "step": 2650 - }, - { - "epoch": 27.203579418344518, - "grad_norm": 0.5198752880096436, - "learning_rate": 1.4670833333333334e-05, - "loss": 0.0472, - "step": 2660 - }, - { - "epoch": 27.305848513902205, - "grad_norm": 0.49260616302490234, - "learning_rate": 1.4650000000000002e-05, - "loss": 0.047, - "step": 2670 - }, - { - "epoch": 27.408117609459893, - "grad_norm": 0.5261008739471436, - "learning_rate": 1.4629166666666667e-05, - "loss": 0.0468, - "step": 2680 - }, - { - "epoch": 27.510386705017577, - "grad_norm": 0.5464609265327454, - "learning_rate": 1.4608333333333335e-05, - "loss": 0.0476, - "step": 2690 - }, - { - "epoch": 27.612655800575265, - "grad_norm": 0.526250422000885, - "learning_rate": 1.45875e-05, - "loss": 0.0488, - "step": 2700 - }, - { - "epoch": 27.612655800575265, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.963590145111084, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8709, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.827, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.978, - "step": 2700 - }, - { - "epoch": 27.71492489613295, - "grad_norm": 0.49739035964012146, - "learning_rate": 1.4566666666666669e-05, - "loss": 0.0478, - "step": 2710 - }, - { - "epoch": 27.817193991690637, - "grad_norm": 0.5307214260101318, - "learning_rate": 1.4545833333333334e-05, - "loss": 0.048, - "step": 2720 - }, - { - "epoch": 27.91946308724832, - "grad_norm": 0.5277805328369141, - "learning_rate": 1.4525e-05, - "loss": 0.0489, - "step": 2730 - }, - { - "epoch": 28.02173218280601, - "grad_norm": 0.490689218044281, - "learning_rate": 1.4504166666666667e-05, - "loss": 0.0488, - "step": 2740 - }, - { - "epoch": 28.124001278363693, - "grad_norm": 0.5121079683303833, - "learning_rate": 1.4483333333333334e-05, - "loss": 0.0426, - "step": 2750 - }, - { - "epoch": 28.22627037392138, - "grad_norm": 0.5197770595550537, - "learning_rate": 1.4462500000000001e-05, - "loss": 0.0433, - "step": 2760 - }, - { - "epoch": 28.328539469479068, - "grad_norm": 0.5023459196090698, - "learning_rate": 1.4441666666666668e-05, - "loss": 0.0428, - "step": 2770 - }, - { - "epoch": 28.430808565036752, - "grad_norm": 0.47794410586357117, - "learning_rate": 1.4420833333333334e-05, - "loss": 0.0431, - "step": 2780 - }, - { - "epoch": 28.53307766059444, - "grad_norm": 0.4780057370662689, - "learning_rate": 1.4400000000000001e-05, - "loss": 0.0439, - "step": 2790 - }, - { - "epoch": 28.635346756152124, - "grad_norm": 0.5096541047096252, - "learning_rate": 1.4379166666666668e-05, - "loss": 0.0443, - "step": 2800 - }, - { - "epoch": 28.635346756152124, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.9908784627914429, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8556, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.897, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.987, - "step": 2800 - }, - { - "epoch": 28.737615851709812, - "grad_norm": 0.49939730763435364, - "learning_rate": 1.4358333333333334e-05, - "loss": 0.045, - "step": 2810 - }, - { - "epoch": 28.839884947267496, - "grad_norm": 0.5106296539306641, - "learning_rate": 1.4337500000000001e-05, - "loss": 0.045, - "step": 2820 - }, - { - "epoch": 28.942154042825184, - "grad_norm": 0.4849531054496765, - "learning_rate": 1.4316666666666668e-05, - "loss": 0.0451, - "step": 2830 - }, - { - "epoch": 29.04442313838287, - "grad_norm": 0.46746712923049927, - "learning_rate": 1.4295833333333335e-05, - "loss": 0.0429, - "step": 2840 - }, - { - "epoch": 29.146692233940556, - "grad_norm": 0.4314974546432495, - "learning_rate": 1.4275000000000001e-05, - "loss": 0.0397, - "step": 2850 - }, - { - "epoch": 29.248961329498243, - "grad_norm": 0.4507407546043396, - "learning_rate": 1.4254166666666668e-05, - "loss": 0.0394, - "step": 2860 - }, - { - "epoch": 29.351230425055927, - "grad_norm": 0.4565179646015167, - "learning_rate": 1.4233333333333335e-05, - "loss": 0.0402, - "step": 2870 - }, - { - "epoch": 29.453499520613615, - "grad_norm": 0.4820208251476288, - "learning_rate": 1.42125e-05, - "loss": 0.0403, - "step": 2880 - }, - { - "epoch": 29.5557686161713, - "grad_norm": 0.47718533873558044, - "learning_rate": 1.4191666666666668e-05, - "loss": 0.0404, - "step": 2890 - }, - { - "epoch": 29.658037711728987, - "grad_norm": 0.5102068185806274, - "learning_rate": 1.4170833333333333e-05, - "loss": 0.0418, - "step": 2900 - }, - { - "epoch": 29.658037711728987, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.013613700866699, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8531, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.909, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.989, - "step": 2900 - }, - { - "epoch": 29.760306807286675, - "grad_norm": 0.49850982427597046, - "learning_rate": 1.4150000000000002e-05, - "loss": 0.041, - "step": 2910 - }, - { - "epoch": 29.86257590284436, - "grad_norm": 0.5028281211853027, - "learning_rate": 1.4129166666666667e-05, - "loss": 0.0419, - "step": 2920 - }, - { - "epoch": 29.964844998402047, - "grad_norm": 0.4753844141960144, - "learning_rate": 1.4108333333333335e-05, - "loss": 0.0421, - "step": 2930 - }, - { - "epoch": 30.06711409395973, - "grad_norm": 0.45387542247772217, - "learning_rate": 1.40875e-05, - "loss": 0.0381, - "step": 2940 - }, - { - "epoch": 30.16938318951742, - "grad_norm": 0.4576801359653473, - "learning_rate": 1.4066666666666669e-05, - "loss": 0.0365, - "step": 2950 - }, - { - "epoch": 30.271652285075103, - "grad_norm": 1.5916332006454468, - "learning_rate": 1.4045833333333334e-05, - "loss": 0.0371, - "step": 2960 - }, - { - "epoch": 30.37392138063279, - "grad_norm": 28.43450164794922, - "learning_rate": 1.4029166666666668e-05, - "loss": 0.0403, - "step": 2970 - }, - { - "epoch": 30.476190476190474, - "grad_norm": 23.308544158935547, - "learning_rate": 1.4010416666666669e-05, - "loss": 0.3001, - "step": 2980 - }, - { - "epoch": 30.578459571748162, - "grad_norm": 0.9334068298339844, - "learning_rate": 1.3989583333333334e-05, - "loss": 0.2114, - "step": 2990 - }, - { - "epoch": 30.68072866730585, - "grad_norm": 0.5433509945869446, - "learning_rate": 1.3968750000000002e-05, - "loss": 0.0448, - "step": 3000 - }, - { - "epoch": 30.68072866730585, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.014845609664917, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8553, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.899, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.987, - "step": 3000 - }, - { - "epoch": 30.782997762863534, - "grad_norm": 0.5084338188171387, - "learning_rate": 1.3947916666666667e-05, - "loss": 0.042, - "step": 3010 - }, - { - "epoch": 30.88526685842122, - "grad_norm": 0.4836946725845337, - "learning_rate": 1.3927083333333336e-05, - "loss": 0.0396, - "step": 3020 - }, - { - "epoch": 30.987535953978906, - "grad_norm": 0.4835638701915741, - "learning_rate": 1.3906250000000001e-05, - "loss": 0.0397, - "step": 3030 - }, - { - "epoch": 31.089805049536594, - "grad_norm": 0.4484618008136749, - "learning_rate": 1.3885416666666666e-05, - "loss": 0.0359, - "step": 3040 - }, - { - "epoch": 31.192074145094278, - "grad_norm": 0.4576917886734009, - "learning_rate": 1.3864583333333334e-05, - "loss": 0.0352, - "step": 3050 - }, - { - "epoch": 31.294343240651965, - "grad_norm": 0.4618643820285797, - "learning_rate": 1.3843750000000001e-05, - "loss": 0.0351, - "step": 3060 - }, - { - "epoch": 31.396612336209653, - "grad_norm": 0.4532334804534912, - "learning_rate": 1.3822916666666668e-05, - "loss": 0.0353, - "step": 3070 - }, - { - "epoch": 31.498881431767337, - "grad_norm": 0.43523749709129333, - "learning_rate": 1.3802083333333335e-05, - "loss": 0.0357, - "step": 3080 - }, - { - "epoch": 31.601150527325025, - "grad_norm": 0.4621034860610962, - "learning_rate": 1.3781250000000001e-05, - "loss": 0.036, - "step": 3090 - }, - { - "epoch": 31.70341962288271, - "grad_norm": 0.4407444894313812, - "learning_rate": 1.3760416666666668e-05, - "loss": 0.0365, - "step": 3100 - }, - { - "epoch": 31.70341962288271, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.0446879863739014, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8565, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.893, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.987, - "step": 3100 - }, - { - "epoch": 31.805688718440397, - "grad_norm": 0.46584275364875793, - "learning_rate": 1.3739583333333335e-05, - "loss": 0.0367, - "step": 3110 - }, - { - "epoch": 31.90795781399808, - "grad_norm": 0.4602925181388855, - "learning_rate": 1.3718750000000001e-05, - "loss": 0.0368, - "step": 3120 - }, - { - "epoch": 32.01022690955577, - "grad_norm": 0.4346022605895996, - "learning_rate": 1.3697916666666668e-05, - "loss": 0.0398, - "step": 3130 - }, - { - "epoch": 32.11249600511346, - "grad_norm": 0.42625728249549866, - "learning_rate": 1.3677083333333335e-05, - "loss": 0.0327, - "step": 3140 - }, - { - "epoch": 32.214765100671144, - "grad_norm": 0.4581526219844818, - "learning_rate": 1.3656250000000002e-05, - "loss": 0.0325, - "step": 3150 - }, - { - "epoch": 32.317034196228825, - "grad_norm": 0.41788893938064575, - "learning_rate": 1.3635416666666668e-05, - "loss": 0.0323, - "step": 3160 - }, - { - "epoch": 32.41930329178651, - "grad_norm": 0.39856305718421936, - "learning_rate": 1.3614583333333335e-05, - "loss": 0.0331, - "step": 3170 - }, - { - "epoch": 32.5215723873442, - "grad_norm": 0.4417785704135895, - "learning_rate": 1.3593750000000002e-05, - "loss": 0.0336, - "step": 3180 - }, - { - "epoch": 32.62384148290189, - "grad_norm": 0.45321398973464966, - "learning_rate": 1.3572916666666667e-05, - "loss": 0.0339, - "step": 3190 - }, - { - "epoch": 32.72611057845957, - "grad_norm": 0.43776148557662964, - "learning_rate": 1.3552083333333335e-05, - "loss": 0.0337, - "step": 3200 - }, - { - "epoch": 32.72611057845957, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.0593974590301514, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8377, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.98, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.997, - "step": 3200 - }, - { - "epoch": 32.828379674017256, - "grad_norm": 0.4577961564064026, - "learning_rate": 1.353125e-05, - "loss": 0.0342, - "step": 3210 - }, - { - "epoch": 32.930648769574944, - "grad_norm": 0.43651047348976135, - "learning_rate": 1.3510416666666669e-05, - "loss": 0.0352, - "step": 3220 - }, - { - "epoch": 33.03291786513263, - "grad_norm": 0.4300122559070587, - "learning_rate": 1.3489583333333334e-05, - "loss": 0.0331, - "step": 3230 - }, - { - "epoch": 33.13518696069032, - "grad_norm": 0.43906641006469727, - "learning_rate": 1.3468750000000002e-05, - "loss": 0.0304, - "step": 3240 - }, - { - "epoch": 33.237456056248, - "grad_norm": 0.40852677822113037, - "learning_rate": 1.3447916666666667e-05, - "loss": 0.0304, - "step": 3250 - }, - { - "epoch": 33.33972515180569, - "grad_norm": 0.42338284850120544, - "learning_rate": 1.3427083333333336e-05, - "loss": 0.0307, - "step": 3260 - }, - { - "epoch": 33.441994247363375, - "grad_norm": 0.4441679120063782, - "learning_rate": 1.340625e-05, - "loss": 0.031, - "step": 3270 - }, - { - "epoch": 33.54426334292106, - "grad_norm": 0.431755393743515, - "learning_rate": 1.3385416666666669e-05, - "loss": 0.031, - "step": 3280 - }, - { - "epoch": 33.646532438478744, - "grad_norm": 0.42783161997795105, - "learning_rate": 1.3364583333333334e-05, - "loss": 0.0316, - "step": 3290 - }, - { - "epoch": 33.74880153403643, - "grad_norm": 0.4471502900123596, - "learning_rate": 1.3343750000000002e-05, - "loss": 0.0318, - "step": 3300 - }, - { - "epoch": 33.74880153403643, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.0793018341064453, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.9135, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.632, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.954, - "step": 3300 - }, - { - "epoch": 33.85107062959412, - "grad_norm": 0.42543718218803406, - "learning_rate": 1.3322916666666668e-05, - "loss": 0.0317, - "step": 3310 - }, - { - "epoch": 33.95333972515181, - "grad_norm": 0.4455837905406952, - "learning_rate": 1.3302083333333336e-05, - "loss": 0.0321, - "step": 3320 - }, - { - "epoch": 34.055608820709494, - "grad_norm": 0.451027512550354, - "learning_rate": 1.3281250000000001e-05, - "loss": 0.0297, - "step": 3330 - }, - { - "epoch": 34.157877916267175, - "grad_norm": 0.43288454413414, - "learning_rate": 1.3260416666666666e-05, - "loss": 0.0285, - "step": 3340 - }, - { - "epoch": 34.26014701182486, - "grad_norm": 0.41506046056747437, - "learning_rate": 1.3239583333333334e-05, - "loss": 0.0285, - "step": 3350 - }, - { - "epoch": 34.36241610738255, - "grad_norm": 0.4198153614997864, - "learning_rate": 1.321875e-05, - "loss": 0.0287, - "step": 3360 - }, - { - "epoch": 34.46468520294024, - "grad_norm": 0.4576322138309479, - "learning_rate": 1.3197916666666668e-05, - "loss": 0.0293, - "step": 3370 - }, - { - "epoch": 34.566954298497926, - "grad_norm": 0.4268178343772888, - "learning_rate": 1.3177083333333333e-05, - "loss": 0.0297, - "step": 3380 - }, - { - "epoch": 34.66922339405561, - "grad_norm": 0.4406622648239136, - "learning_rate": 1.3156250000000001e-05, - "loss": 0.0298, - "step": 3390 - }, - { - "epoch": 34.771492489613294, - "grad_norm": 0.44503384828567505, - "learning_rate": 1.3135416666666666e-05, - "loss": 0.0302, - "step": 3400 - }, - { - "epoch": 34.771492489613294, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1051576137542725, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7904, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.198, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.025, - "step": 3400 - }, - { - "epoch": 34.87376158517098, - "grad_norm": 0.4410327672958374, - "learning_rate": 1.3114583333333335e-05, - "loss": 0.0308, - "step": 3410 - }, - { - "epoch": 34.97603068072867, - "grad_norm": 0.4527595639228821, - "learning_rate": 1.309375e-05, - "loss": 0.0307, - "step": 3420 - }, - { - "epoch": 35.07829977628635, - "grad_norm": 0.3983278274536133, - "learning_rate": 1.3072916666666668e-05, - "loss": 0.0275, - "step": 3430 - }, - { - "epoch": 35.18056887184404, - "grad_norm": 0.405274361371994, - "learning_rate": 1.3052083333333335e-05, - "loss": 0.027, - "step": 3440 - }, - { - "epoch": 35.282837967401726, - "grad_norm": 0.4225080907344818, - "learning_rate": 1.3031250000000002e-05, - "loss": 0.0276, - "step": 3450 - }, - { - "epoch": 35.38510706295941, - "grad_norm": 0.3988070487976074, - "learning_rate": 1.3010416666666668e-05, - "loss": 0.0271, - "step": 3460 - }, - { - "epoch": 35.4873761585171, - "grad_norm": 0.613850474357605, - "learning_rate": 1.2989583333333335e-05, - "loss": 0.0271, - "step": 3470 - }, - { - "epoch": 35.58964525407478, - "grad_norm": 0.42930155992507935, - "learning_rate": 1.2968750000000002e-05, - "loss": 0.0269, - "step": 3480 - }, - { - "epoch": 35.69191434963247, - "grad_norm": 0.4209059178829193, - "learning_rate": 1.2947916666666667e-05, - "loss": 0.0281, - "step": 3490 - }, - { - "epoch": 35.79418344519016, - "grad_norm": 0.4427293837070465, - "learning_rate": 1.2927083333333335e-05, - "loss": 0.0282, - "step": 3500 - }, - { - "epoch": 35.79418344519016, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1160192489624023, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8212, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.056, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.007, - "step": 3500 - }, - { - "epoch": 35.896452540747845, - "grad_norm": 0.42366185784339905, - "learning_rate": 1.290625e-05, - "loss": 0.0289, - "step": 3510 - }, - { - "epoch": 35.99872163630553, - "grad_norm": 0.4120742380619049, - "learning_rate": 1.2885416666666669e-05, - "loss": 0.0287, - "step": 3520 - }, - { - "epoch": 36.10099073186321, - "grad_norm": 0.38890495896339417, - "learning_rate": 1.2864583333333334e-05, - "loss": 0.0258, - "step": 3530 - }, - { - "epoch": 36.2032598274209, - "grad_norm": 0.4076452851295471, - "learning_rate": 1.2843750000000002e-05, - "loss": 0.0258, - "step": 3540 - }, - { - "epoch": 36.30552892297859, - "grad_norm": 0.3809320032596588, - "learning_rate": 1.2822916666666667e-05, - "loss": 0.0255, - "step": 3550 - }, - { - "epoch": 36.407798018536276, - "grad_norm": 0.39650237560272217, - "learning_rate": 1.2802083333333336e-05, - "loss": 0.026, - "step": 3560 - }, - { - "epoch": 36.51006711409396, - "grad_norm": 0.4009132385253906, - "learning_rate": 1.278125e-05, - "loss": 0.0265, - "step": 3570 - }, - { - "epoch": 36.612336209651644, - "grad_norm": 0.42308393120765686, - "learning_rate": 1.2760416666666669e-05, - "loss": 0.0263, - "step": 3580 - }, - { - "epoch": 36.71460530520933, - "grad_norm": 0.4044691324234009, - "learning_rate": 1.2739583333333334e-05, - "loss": 0.0265, - "step": 3590 - }, - { - "epoch": 36.81687440076702, - "grad_norm": 0.4140937030315399, - "learning_rate": 1.2718750000000003e-05, - "loss": 0.0266, - "step": 3600 - }, - { - "epoch": 36.81687440076702, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.13891339302063, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8236, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.045, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.006, - "step": 3600 - }, - { - "epoch": 36.91914349632471, - "grad_norm": 0.4160206615924835, - "learning_rate": 1.2697916666666668e-05, - "loss": 0.0269, - "step": 3610 - }, - { - "epoch": 37.02141259188239, - "grad_norm": 0.39423078298568726, - "learning_rate": 1.2677083333333336e-05, - "loss": 0.026, - "step": 3620 - }, - { - "epoch": 37.123681687440076, - "grad_norm": 0.3859294056892395, - "learning_rate": 1.2656250000000001e-05, - "loss": 0.0238, - "step": 3630 - }, - { - "epoch": 37.225950782997764, - "grad_norm": 0.40413331985473633, - "learning_rate": 1.2635416666666666e-05, - "loss": 0.0244, - "step": 3640 - }, - { - "epoch": 37.32821987855545, - "grad_norm": 0.39002159237861633, - "learning_rate": 1.2614583333333334e-05, - "loss": 0.0244, - "step": 3650 - }, - { - "epoch": 37.43048897411313, - "grad_norm": 0.403145432472229, - "learning_rate": 1.259375e-05, - "loss": 0.0245, - "step": 3660 - }, - { - "epoch": 37.53275806967082, - "grad_norm": 0.42878827452659607, - "learning_rate": 1.2572916666666668e-05, - "loss": 0.0249, - "step": 3670 - }, - { - "epoch": 37.63502716522851, - "grad_norm": 0.4047834277153015, - "learning_rate": 1.2552083333333333e-05, - "loss": 0.0246, - "step": 3680 - }, - { - "epoch": 37.737296260786195, - "grad_norm": 0.4242531955242157, - "learning_rate": 1.2531250000000001e-05, - "loss": 0.0254, - "step": 3690 - }, - { - "epoch": 37.83956535634388, - "grad_norm": 0.3883196711540222, - "learning_rate": 1.2510416666666666e-05, - "loss": 0.025, - "step": 3700 - }, - { - "epoch": 37.83956535634388, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.135497570037842, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.828, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.024, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.003, - "step": 3700 - }, - { - "epoch": 37.94183445190156, - "grad_norm": 0.40664049983024597, - "learning_rate": 1.2489583333333335e-05, - "loss": 0.0249, - "step": 3710 - }, - { - "epoch": 38.04410354745925, - "grad_norm": 0.37315633893013, - "learning_rate": 1.246875e-05, - "loss": 0.0242, - "step": 3720 - }, - { - "epoch": 38.14637264301694, - "grad_norm": 0.37339115142822266, - "learning_rate": 1.2447916666666668e-05, - "loss": 0.0229, - "step": 3730 - }, - { - "epoch": 38.24864173857463, - "grad_norm": 0.38532310724258423, - "learning_rate": 1.2427083333333333e-05, - "loss": 0.0227, - "step": 3740 - }, - { - "epoch": 38.350910834132314, - "grad_norm": 0.3730473220348358, - "learning_rate": 1.2406250000000002e-05, - "loss": 0.0227, - "step": 3750 - }, - { - "epoch": 38.453179929689995, - "grad_norm": 0.37635692954063416, - "learning_rate": 1.2385416666666667e-05, - "loss": 0.0234, - "step": 3760 - }, - { - "epoch": 38.55544902524768, - "grad_norm": 0.38997572660446167, - "learning_rate": 1.2364583333333335e-05, - "loss": 0.0233, - "step": 3770 - }, - { - "epoch": 38.65771812080537, - "grad_norm": 0.38781270384788513, - "learning_rate": 1.234375e-05, - "loss": 0.0237, - "step": 3780 - }, - { - "epoch": 38.75998721636306, - "grad_norm": 0.42525115609169006, - "learning_rate": 1.2322916666666667e-05, - "loss": 0.0237, - "step": 3790 - }, - { - "epoch": 38.86225631192074, - "grad_norm": 0.3896453380584717, - "learning_rate": 1.2302083333333335e-05, - "loss": 0.0239, - "step": 3800 - }, - { - "epoch": 38.86225631192074, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.166919231414795, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8189, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.067, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.008, - "step": 3800 - }, - { - "epoch": 38.964525407478426, - "grad_norm": 0.3976113498210907, - "learning_rate": 1.228125e-05, - "loss": 0.0242, - "step": 3810 - }, - { - "epoch": 39.066794503036114, - "grad_norm": 0.37402892112731934, - "learning_rate": 1.2260416666666669e-05, - "loss": 0.0228, - "step": 3820 - }, - { - "epoch": 39.1690635985938, - "grad_norm": 0.4018329679965973, - "learning_rate": 1.2239583333333334e-05, - "loss": 0.0219, - "step": 3830 - }, - { - "epoch": 39.27133269415149, - "grad_norm": 0.3905806839466095, - "learning_rate": 1.2218750000000002e-05, - "loss": 0.0219, - "step": 3840 - }, - { - "epoch": 39.37360178970917, - "grad_norm": 0.39368128776550293, - "learning_rate": 1.2197916666666667e-05, - "loss": 0.0223, - "step": 3850 - }, - { - "epoch": 39.47587088526686, - "grad_norm": 0.40275242924690247, - "learning_rate": 1.2177083333333336e-05, - "loss": 0.0225, - "step": 3860 - }, - { - "epoch": 39.578139980824545, - "grad_norm": 0.377655953168869, - "learning_rate": 1.215625e-05, - "loss": 0.0223, - "step": 3870 - }, - { - "epoch": 39.68040907638223, - "grad_norm": 0.35910850763320923, - "learning_rate": 1.2135416666666669e-05, - "loss": 0.0222, - "step": 3880 - }, - { - "epoch": 39.782678171939914, - "grad_norm": 0.40253758430480957, - "learning_rate": 1.2114583333333334e-05, - "loss": 0.0229, - "step": 3890 - }, - { - "epoch": 39.8849472674976, - "grad_norm": 0.4119781255722046, - "learning_rate": 1.2093750000000003e-05, - "loss": 0.0231, - "step": 3900 - }, - { - "epoch": 39.8849472674976, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1834068298339844, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8776, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.796, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.975, - "step": 3900 - }, - { - "epoch": 39.98721636305529, - "grad_norm": 0.40584734082221985, - "learning_rate": 1.2072916666666668e-05, - "loss": 0.0229, - "step": 3910 - }, - { - "epoch": 40.08948545861298, - "grad_norm": 0.3695628345012665, - "learning_rate": 1.2052083333333336e-05, - "loss": 0.0207, - "step": 3920 - }, - { - "epoch": 40.191754554170664, - "grad_norm": 0.38678476214408875, - "learning_rate": 1.2031250000000001e-05, - "loss": 0.0205, - "step": 3930 - }, - { - "epoch": 40.294023649728345, - "grad_norm": 0.349589079618454, - "learning_rate": 1.2010416666666666e-05, - "loss": 0.0208, - "step": 3940 - }, - { - "epoch": 40.39629274528603, - "grad_norm": 0.3602575659751892, - "learning_rate": 1.1989583333333335e-05, - "loss": 0.0207, - "step": 3950 - }, - { - "epoch": 40.49856184084372, - "grad_norm": 0.34940171241760254, - "learning_rate": 1.196875e-05, - "loss": 0.0208, - "step": 3960 - }, - { - "epoch": 40.60083093640141, - "grad_norm": 0.42674344778060913, - "learning_rate": 1.1947916666666668e-05, - "loss": 0.0214, - "step": 3970 - }, - { - "epoch": 40.703100031959096, - "grad_norm": 0.3709782361984253, - "learning_rate": 1.1927083333333333e-05, - "loss": 0.0215, - "step": 3980 - }, - { - "epoch": 40.80536912751678, - "grad_norm": 0.37002718448638916, - "learning_rate": 1.1906250000000001e-05, - "loss": 0.0215, - "step": 3990 - }, - { - "epoch": 40.907638223074464, - "grad_norm": 0.3808917999267578, - "learning_rate": 1.1885416666666666e-05, - "loss": 0.0216, - "step": 4000 - }, - { - "epoch": 40.907638223074464, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.173332452774048, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8356, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.99, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.999, - "step": 4000 - }, - { - "epoch": 41.00990731863215, - "grad_norm": 0.5391054153442383, - "learning_rate": 1.1864583333333335e-05, - "loss": 0.0214, - "step": 4010 - }, - { - "epoch": 41.11217641418984, - "grad_norm": 0.36861568689346313, - "learning_rate": 1.184375e-05, - "loss": 0.0203, - "step": 4020 - }, - { - "epoch": 41.21444550974752, - "grad_norm": 0.3627229630947113, - "learning_rate": 1.1822916666666668e-05, - "loss": 0.0203, - "step": 4030 - }, - { - "epoch": 41.31671460530521, - "grad_norm": 0.37450557947158813, - "learning_rate": 1.1802083333333333e-05, - "loss": 0.02, - "step": 4040 - }, - { - "epoch": 41.418983700862896, - "grad_norm": 0.3615454435348511, - "learning_rate": 1.1781250000000002e-05, - "loss": 0.0204, - "step": 4050 - }, - { - "epoch": 41.52125279642058, - "grad_norm": 0.3542068302631378, - "learning_rate": 1.1760416666666667e-05, - "loss": 0.0203, - "step": 4060 - }, - { - "epoch": 41.62352189197827, - "grad_norm": 0.3920552134513855, - "learning_rate": 1.1739583333333335e-05, - "loss": 0.0204, - "step": 4070 - }, - { - "epoch": 41.72579098753595, - "grad_norm": 0.3835306763648987, - "learning_rate": 1.171875e-05, - "loss": 0.0206, - "step": 4080 - }, - { - "epoch": 41.82806008309364, - "grad_norm": 0.35131242871284485, - "learning_rate": 1.1697916666666667e-05, - "loss": 0.0208, - "step": 4090 - }, - { - "epoch": 41.93032917865133, - "grad_norm": 0.3543274402618408, - "learning_rate": 1.1677083333333334e-05, - "loss": 0.0202, - "step": 4100 - }, - { - "epoch": 41.93032917865133, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1988096237182617, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8148, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.086, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.011, - "step": 4100 - }, - { - "epoch": 42.032598274209015, - "grad_norm": 0.33442866802215576, - "learning_rate": 1.165625e-05, - "loss": 0.0198, - "step": 4110 - }, - { - "epoch": 42.134867369766695, - "grad_norm": 0.3324146866798401, - "learning_rate": 1.1635416666666667e-05, - "loss": 0.0192, - "step": 4120 - }, - { - "epoch": 42.23713646532438, - "grad_norm": 0.3776349723339081, - "learning_rate": 1.1614583333333334e-05, - "loss": 0.0191, - "step": 4130 - }, - { - "epoch": 42.33940556088207, - "grad_norm": 0.3571579158306122, - "learning_rate": 1.159375e-05, - "loss": 0.0192, - "step": 4140 - }, - { - "epoch": 42.44167465643976, - "grad_norm": 0.37327298521995544, - "learning_rate": 1.1572916666666667e-05, - "loss": 0.0192, - "step": 4150 - }, - { - "epoch": 42.543943751997446, - "grad_norm": 0.3787640631198883, - "learning_rate": 1.1552083333333334e-05, - "loss": 0.0196, - "step": 4160 - }, - { - "epoch": 42.64621284755513, - "grad_norm": 0.36109158396720886, - "learning_rate": 1.153125e-05, - "loss": 0.0193, - "step": 4170 - }, - { - "epoch": 42.748481943112814, - "grad_norm": 0.34873998165130615, - "learning_rate": 1.151041666666667e-05, - "loss": 0.0194, - "step": 4180 - }, - { - "epoch": 42.8507510386705, - "grad_norm": 0.393928587436676, - "learning_rate": 1.1489583333333334e-05, - "loss": 0.0196, - "step": 4190 - }, - { - "epoch": 42.95302013422819, - "grad_norm": 0.3687518537044525, - "learning_rate": 1.1468750000000003e-05, - "loss": 0.0198, - "step": 4200 - }, - { - "epoch": 42.95302013422819, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2081563472747803, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7395, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.434, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.054, - "step": 4200 - }, - { - "epoch": 43.05528922978588, - "grad_norm": 0.3610316514968872, - "learning_rate": 1.1447916666666668e-05, - "loss": 0.0187, - "step": 4210 - }, - { - "epoch": 43.15755832534356, - "grad_norm": 0.3438417911529541, - "learning_rate": 1.1427083333333336e-05, - "loss": 0.0181, - "step": 4220 - }, - { - "epoch": 43.259827420901246, - "grad_norm": 0.33607029914855957, - "learning_rate": 1.1406250000000001e-05, - "loss": 0.018, - "step": 4230 - }, - { - "epoch": 43.362096516458934, - "grad_norm": 0.3655332326889038, - "learning_rate": 1.1385416666666666e-05, - "loss": 0.018, - "step": 4240 - }, - { - "epoch": 43.46436561201662, - "grad_norm": 0.3450184762477875, - "learning_rate": 1.1364583333333335e-05, - "loss": 0.0182, - "step": 4250 - }, - { - "epoch": 43.5666347075743, - "grad_norm": 0.34371188282966614, - "learning_rate": 1.134375e-05, - "loss": 0.0185, - "step": 4260 - }, - { - "epoch": 43.66890380313199, - "grad_norm": 0.3620694875717163, - "learning_rate": 1.1322916666666668e-05, - "loss": 0.0188, - "step": 4270 - }, - { - "epoch": 43.77117289868968, - "grad_norm": 0.34563344717025757, - "learning_rate": 1.1302083333333333e-05, - "loss": 0.0185, - "step": 4280 - }, - { - "epoch": 43.873441994247365, - "grad_norm": 0.3630271255970001, - "learning_rate": 1.1281250000000001e-05, - "loss": 0.0189, - "step": 4290 - }, - { - "epoch": 43.97571108980505, - "grad_norm": 0.3588694930076599, - "learning_rate": 1.1260416666666666e-05, - "loss": 0.0192, - "step": 4300 - }, - { - "epoch": 43.97571108980505, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.21140456199646, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7918, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.192, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.024, - "step": 4300 - }, - { - "epoch": 44.07798018536273, - "grad_norm": 0.3162562847137451, - "learning_rate": 1.1239583333333335e-05, - "loss": 0.0172, - "step": 4310 - }, - { - "epoch": 44.18024928092042, - "grad_norm": 0.366394966840744, - "learning_rate": 1.121875e-05, - "loss": 0.017, - "step": 4320 - }, - { - "epoch": 44.28251837647811, - "grad_norm": 0.35083746910095215, - "learning_rate": 1.1197916666666668e-05, - "loss": 0.0178, - "step": 4330 - }, - { - "epoch": 44.384787472035796, - "grad_norm": 0.35160747170448303, - "learning_rate": 1.1177083333333333e-05, - "loss": 0.0174, - "step": 4340 - }, - { - "epoch": 44.48705656759348, - "grad_norm": 0.35368478298187256, - "learning_rate": 1.1156250000000002e-05, - "loss": 0.0175, - "step": 4350 - }, - { - "epoch": 44.589325663151165, - "grad_norm": 0.3595126271247864, - "learning_rate": 1.1135416666666667e-05, - "loss": 0.0176, - "step": 4360 - }, - { - "epoch": 44.69159475870885, - "grad_norm": 0.34251803159713745, - "learning_rate": 1.1114583333333335e-05, - "loss": 0.0179, - "step": 4370 - }, - { - "epoch": 44.79386385426654, - "grad_norm": 0.3488711416721344, - "learning_rate": 1.109375e-05, - "loss": 0.018, - "step": 4380 - }, - { - "epoch": 44.89613294982423, - "grad_norm": 0.35304006934165955, - "learning_rate": 1.1072916666666667e-05, - "loss": 0.018, - "step": 4390 - }, - { - "epoch": 44.99840204538191, - "grad_norm": 0.3306570053100586, - "learning_rate": 1.1052083333333334e-05, - "loss": 0.018, - "step": 4400 - }, - { - "epoch": 44.99840204538191, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.226471424102783, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7401, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.432, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.054, - "step": 4400 - }, - { - "epoch": 45.100671140939596, - "grad_norm": 0.3414161801338196, - "learning_rate": 1.103125e-05, - "loss": 0.0163, - "step": 4410 - }, - { - "epoch": 45.202940236497284, - "grad_norm": 0.34503790736198425, - "learning_rate": 1.1010416666666667e-05, - "loss": 0.0167, - "step": 4420 - }, - { - "epoch": 45.30520933205497, - "grad_norm": 0.3578576147556305, - "learning_rate": 1.0989583333333334e-05, - "loss": 0.0165, - "step": 4430 - }, - { - "epoch": 45.40747842761266, - "grad_norm": 0.34795865416526794, - "learning_rate": 1.096875e-05, - "loss": 0.0165, - "step": 4440 - }, - { - "epoch": 45.50974752317034, - "grad_norm": 0.35079431533813477, - "learning_rate": 1.0947916666666667e-05, - "loss": 0.0168, - "step": 4450 - }, - { - "epoch": 45.61201661872803, - "grad_norm": 0.30789047479629517, - "learning_rate": 1.0927083333333334e-05, - "loss": 0.0164, - "step": 4460 - }, - { - "epoch": 45.714285714285715, - "grad_norm": 0.370721697807312, - "learning_rate": 1.090625e-05, - "loss": 0.0174, - "step": 4470 - }, - { - "epoch": 45.8165548098434, - "grad_norm": 0.34695321321487427, - "learning_rate": 1.0885416666666668e-05, - "loss": 0.0171, - "step": 4480 - }, - { - "epoch": 45.918823905401084, - "grad_norm": 0.3586573600769043, - "learning_rate": 1.0864583333333334e-05, - "loss": 0.0171, - "step": 4490 - }, - { - "epoch": 46.02109300095877, - "grad_norm": 0.3352271318435669, - "learning_rate": 1.0843750000000001e-05, - "loss": 0.0172, - "step": 4500 - }, - { - "epoch": 46.02109300095877, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2414653301239014, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.694, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.646, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.081, - "step": 4500 - }, - { - "epoch": 46.12336209651646, - "grad_norm": 0.3106061518192291, - "learning_rate": 1.0822916666666668e-05, - "loss": 0.0154, - "step": 4510 - }, - { - "epoch": 46.22563119207415, - "grad_norm": 0.3251485824584961, - "learning_rate": 1.0802083333333334e-05, - "loss": 0.0159, - "step": 4520 - }, - { - "epoch": 46.327900287631834, - "grad_norm": 0.3339298367500305, - "learning_rate": 1.0781250000000001e-05, - "loss": 0.0157, - "step": 4530 - }, - { - "epoch": 46.430169383189515, - "grad_norm": 0.33735695481300354, - "learning_rate": 1.0760416666666666e-05, - "loss": 0.0161, - "step": 4540 - }, - { - "epoch": 46.5324384787472, - "grad_norm": 0.3049178719520569, - "learning_rate": 1.0739583333333335e-05, - "loss": 0.0162, - "step": 4550 - }, - { - "epoch": 46.63470757430489, - "grad_norm": 0.3524739742279053, - "learning_rate": 1.071875e-05, - "loss": 0.016, - "step": 4560 - }, - { - "epoch": 46.73697666986258, - "grad_norm": 0.3321375250816345, - "learning_rate": 1.0697916666666668e-05, - "loss": 0.016, - "step": 4570 - }, - { - "epoch": 46.83924576542026, - "grad_norm": 0.3346993029117584, - "learning_rate": 1.0677083333333333e-05, - "loss": 0.0165, - "step": 4580 - }, - { - "epoch": 46.94151486097795, - "grad_norm": 0.3200349807739258, - "learning_rate": 1.0656250000000002e-05, - "loss": 0.0162, - "step": 4590 - }, - { - "epoch": 47.043783956535634, - "grad_norm": 0.3281605541706085, - "learning_rate": 1.0635416666666667e-05, - "loss": 0.0158, - "step": 4600 - }, - { - "epoch": 47.043783956535634, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.256087303161621, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.744, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.413, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.052, - "step": 4600 - }, - { - "epoch": 47.14605305209332, - "grad_norm": 0.35677453875541687, - "learning_rate": 1.0614583333333335e-05, - "loss": 0.0153, - "step": 4610 - }, - { - "epoch": 47.24832214765101, - "grad_norm": 0.2933562099933624, - "learning_rate": 1.059375e-05, - "loss": 0.0151, - "step": 4620 - }, - { - "epoch": 47.35059124320869, - "grad_norm": 0.3412613272666931, - "learning_rate": 1.0572916666666668e-05, - "loss": 0.0154, - "step": 4630 - }, - { - "epoch": 47.45286033876638, - "grad_norm": 0.32954832911491394, - "learning_rate": 1.0552083333333333e-05, - "loss": 0.0154, - "step": 4640 - }, - { - "epoch": 47.555129434324066, - "grad_norm": 0.3282068073749542, - "learning_rate": 1.0531250000000002e-05, - "loss": 0.0152, - "step": 4650 - }, - { - "epoch": 47.65739852988175, - "grad_norm": 0.33588287234306335, - "learning_rate": 1.0510416666666667e-05, - "loss": 0.0156, - "step": 4660 - }, - { - "epoch": 47.75966762543944, - "grad_norm": 0.33600950241088867, - "learning_rate": 1.0489583333333335e-05, - "loss": 0.0151, - "step": 4670 - }, - { - "epoch": 47.86193672099712, - "grad_norm": 0.335001140832901, - "learning_rate": 1.046875e-05, - "loss": 0.0156, - "step": 4680 - }, - { - "epoch": 47.96420581655481, - "grad_norm": 0.33093252778053284, - "learning_rate": 1.0447916666666667e-05, - "loss": 0.0158, - "step": 4690 - }, - { - "epoch": 48.0664749121125, - "grad_norm": 0.32087039947509766, - "learning_rate": 1.0427083333333334e-05, - "loss": 0.0149, - "step": 4700 - }, - { - "epoch": 48.0664749121125, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.274885654449463, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8036, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.137, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.017, - "step": 4700 - }, - { - "epoch": 48.168744007670185, - "grad_norm": 0.31192752718925476, - "learning_rate": 1.040625e-05, - "loss": 0.0145, - "step": 4710 - }, - { - "epoch": 48.271013103227865, - "grad_norm": 0.2866131067276001, - "learning_rate": 1.0385416666666667e-05, - "loss": 0.0144, - "step": 4720 - }, - { - "epoch": 48.37328219878555, - "grad_norm": 0.34419262409210205, - "learning_rate": 1.0364583333333334e-05, - "loss": 0.0145, - "step": 4730 - }, - { - "epoch": 48.47555129434324, - "grad_norm": 0.33133459091186523, - "learning_rate": 1.034375e-05, - "loss": 0.0143, - "step": 4740 - }, - { - "epoch": 48.57782038990093, - "grad_norm": 0.3273964524269104, - "learning_rate": 1.0322916666666667e-05, - "loss": 0.0146, - "step": 4750 - }, - { - "epoch": 48.680089485458616, - "grad_norm": 0.3400149643421173, - "learning_rate": 1.0302083333333334e-05, - "loss": 0.0149, - "step": 4760 - }, - { - "epoch": 48.7823585810163, - "grad_norm": 0.3198685348033905, - "learning_rate": 1.0281250000000001e-05, - "loss": 0.0148, - "step": 4770 - }, - { - "epoch": 48.884627676573984, - "grad_norm": 0.3375503420829773, - "learning_rate": 1.0260416666666668e-05, - "loss": 0.0153, - "step": 4780 - }, - { - "epoch": 48.98689677213167, - "grad_norm": 0.3139088749885559, - "learning_rate": 1.0239583333333334e-05, - "loss": 0.0152, - "step": 4790 - }, - { - "epoch": 49.08916586768936, - "grad_norm": 0.32369062304496765, - "learning_rate": 1.0218750000000001e-05, - "loss": 0.0143, - "step": 4800 - }, - { - "epoch": 49.08916586768936, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2769861221313477, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7984, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.161, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.02, - "step": 4800 - }, - { - "epoch": 49.19143496324704, - "grad_norm": 0.3071221113204956, - "learning_rate": 1.0197916666666668e-05, - "loss": 0.0136, - "step": 4810 - }, - { - "epoch": 49.29370405880473, - "grad_norm": 0.31288015842437744, - "learning_rate": 1.0177083333333335e-05, - "loss": 0.0139, - "step": 4820 - }, - { - "epoch": 49.395973154362416, - "grad_norm": 0.31299805641174316, - "learning_rate": 1.0156250000000001e-05, - "loss": 0.0137, - "step": 4830 - }, - { - "epoch": 49.498242249920104, - "grad_norm": 0.3377828001976013, - "learning_rate": 1.0135416666666666e-05, - "loss": 0.0141, - "step": 4840 - }, - { - "epoch": 49.60051134547779, - "grad_norm": 0.31973740458488464, - "learning_rate": 1.0114583333333335e-05, - "loss": 0.0144, - "step": 4850 - }, - { - "epoch": 49.70278044103547, - "grad_norm": 0.3018786907196045, - "learning_rate": 1.009375e-05, - "loss": 0.0144, - "step": 4860 - }, - { - "epoch": 49.80504953659316, - "grad_norm": 0.3308105766773224, - "learning_rate": 1.0072916666666668e-05, - "loss": 0.0146, - "step": 4870 - }, - { - "epoch": 49.90731863215085, - "grad_norm": 0.3090561032295227, - "learning_rate": 1.0052083333333333e-05, - "loss": 0.0145, - "step": 4880 - }, - { - "epoch": 50.009587727708535, - "grad_norm": 0.33793970942497253, - "learning_rate": 1.0031250000000002e-05, - "loss": 0.0145, - "step": 4890 - }, - { - "epoch": 50.11185682326622, - "grad_norm": 0.3269507884979248, - "learning_rate": 1.0010416666666667e-05, - "loss": 0.0133, - "step": 4900 - }, - { - "epoch": 50.11185682326622, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2938148975372314, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7798, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.247, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.031, - "step": 4900 - }, - { - "epoch": 50.2141259188239, - "grad_norm": 0.30052992701530457, - "learning_rate": 9.989583333333333e-06, - "loss": 0.0136, - "step": 4910 - }, - { - "epoch": 50.31639501438159, - "grad_norm": 0.2977409362792969, - "learning_rate": 9.96875e-06, - "loss": 0.0134, - "step": 4920 - }, - { - "epoch": 50.41866410993928, - "grad_norm": 0.3022303283214569, - "learning_rate": 9.947916666666667e-06, - "loss": 0.0136, - "step": 4930 - }, - { - "epoch": 50.520933205496966, - "grad_norm": 0.2963425815105438, - "learning_rate": 9.927083333333334e-06, - "loss": 0.0137, - "step": 4940 - }, - { - "epoch": 50.62320230105465, - "grad_norm": 0.2862411141395569, - "learning_rate": 9.90625e-06, - "loss": 0.0134, - "step": 4950 - }, - { - "epoch": 50.725471396612335, - "grad_norm": 0.32539355754852295, - "learning_rate": 9.885416666666667e-06, - "loss": 0.0136, - "step": 4960 - }, - { - "epoch": 50.82774049217002, - "grad_norm": 0.30780264735221863, - "learning_rate": 9.864583333333334e-06, - "loss": 0.0138, - "step": 4970 - }, - { - "epoch": 50.93000958772771, - "grad_norm": 0.3183571994304657, - "learning_rate": 9.84375e-06, - "loss": 0.0136, - "step": 4980 - }, - { - "epoch": 51.0322786832854, - "grad_norm": 0.28671231865882874, - "learning_rate": 9.822916666666667e-06, - "loss": 0.0136, - "step": 4990 - }, - { - "epoch": 51.13454777884308, - "grad_norm": 0.2912745177745819, - "learning_rate": 9.802083333333334e-06, - "loss": 0.0126, - "step": 5000 - }, - { - "epoch": 51.13454777884308, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.312565326690674, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8065, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.124, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.015, - "step": 5000 - }, - { - "epoch": 51.236816874400766, - "grad_norm": 0.29751163721084595, - "learning_rate": 9.78125e-06, - "loss": 0.013, - "step": 5010 - }, - { - "epoch": 51.339085969958454, - "grad_norm": 0.30533960461616516, - "learning_rate": 9.760416666666667e-06, - "loss": 0.013, - "step": 5020 - }, - { - "epoch": 51.44135506551614, - "grad_norm": 0.3039548695087433, - "learning_rate": 9.739583333333334e-06, - "loss": 0.0129, - "step": 5030 - }, - { - "epoch": 51.54362416107382, - "grad_norm": 0.295386403799057, - "learning_rate": 9.71875e-06, - "loss": 0.0132, - "step": 5040 - }, - { - "epoch": 51.64589325663151, - "grad_norm": 0.289328932762146, - "learning_rate": 9.697916666666667e-06, - "loss": 0.0133, - "step": 5050 - }, - { - "epoch": 51.7481623521892, - "grad_norm": 0.3070317804813385, - "learning_rate": 9.677083333333334e-06, - "loss": 0.0131, - "step": 5060 - }, - { - "epoch": 51.850431447746885, - "grad_norm": 0.28922444581985474, - "learning_rate": 9.656250000000001e-06, - "loss": 0.0131, - "step": 5070 - }, - { - "epoch": 51.95270054330457, - "grad_norm": 0.3166629672050476, - "learning_rate": 9.635416666666668e-06, - "loss": 0.0132, - "step": 5080 - }, - { - "epoch": 52.054969638862254, - "grad_norm": 0.3101074993610382, - "learning_rate": 9.614583333333334e-06, - "loss": 0.0126, - "step": 5090 - }, - { - "epoch": 52.15723873441994, - "grad_norm": 0.318968266248703, - "learning_rate": 9.593750000000001e-06, - "loss": 0.0127, - "step": 5100 - }, - { - "epoch": 52.15723873441994, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3045780658721924, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8544, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.903, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.988, - "step": 5100 - }, - { - "epoch": 52.25950782997763, - "grad_norm": 0.30769291520118713, - "learning_rate": 9.572916666666668e-06, - "loss": 0.0122, - "step": 5110 - }, - { - "epoch": 52.36177692553532, - "grad_norm": 0.28409647941589355, - "learning_rate": 9.552083333333335e-06, - "loss": 0.0124, - "step": 5120 - }, - { - "epoch": 52.464046021093004, - "grad_norm": 0.3307039737701416, - "learning_rate": 9.531250000000001e-06, - "loss": 0.0128, - "step": 5130 - }, - { - "epoch": 52.566315116650685, - "grad_norm": 0.292473703622818, - "learning_rate": 9.510416666666668e-06, - "loss": 0.0125, - "step": 5140 - }, - { - "epoch": 52.66858421220837, - "grad_norm": 0.3042745888233185, - "learning_rate": 9.489583333333335e-06, - "loss": 0.0126, - "step": 5150 - }, - { - "epoch": 52.77085330776606, - "grad_norm": 0.29861128330230713, - "learning_rate": 9.468750000000001e-06, - "loss": 0.0128, - "step": 5160 - }, - { - "epoch": 52.87312240332375, - "grad_norm": 0.284404993057251, - "learning_rate": 9.447916666666668e-06, - "loss": 0.0129, - "step": 5170 - }, - { - "epoch": 52.97539149888143, - "grad_norm": 0.30096396803855896, - "learning_rate": 9.427083333333335e-06, - "loss": 0.0131, - "step": 5180 - }, - { - "epoch": 53.077660594439116, - "grad_norm": 0.3048815131187439, - "learning_rate": 9.406250000000002e-06, - "loss": 0.012, - "step": 5190 - }, - { - "epoch": 53.179929689996804, - "grad_norm": 0.2931344509124756, - "learning_rate": 9.385416666666668e-06, - "loss": 0.0118, - "step": 5200 - }, - { - "epoch": 53.179929689996804, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.321913480758667, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8669, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.845, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.981, - "step": 5200 - }, - { - "epoch": 53.28219878555449, - "grad_norm": 0.29786059260368347, - "learning_rate": 9.364583333333333e-06, - "loss": 0.0116, - "step": 5210 - }, - { - "epoch": 53.38446788111218, - "grad_norm": 0.2869616150856018, - "learning_rate": 9.34375e-06, - "loss": 0.0119, - "step": 5220 - }, - { - "epoch": 53.48673697666986, - "grad_norm": 0.3017300069332123, - "learning_rate": 9.322916666666667e-06, - "loss": 0.0121, - "step": 5230 - }, - { - "epoch": 53.58900607222755, - "grad_norm": 0.3326238691806793, - "learning_rate": 9.302083333333334e-06, - "loss": 0.0123, - "step": 5240 - }, - { - "epoch": 53.691275167785236, - "grad_norm": 0.30569109320640564, - "learning_rate": 9.28125e-06, - "loss": 0.0125, - "step": 5250 - }, - { - "epoch": 53.79354426334292, - "grad_norm": 0.3061717748641968, - "learning_rate": 9.260416666666667e-06, - "loss": 0.0123, - "step": 5260 - }, - { - "epoch": 53.895813358900604, - "grad_norm": 0.281955748796463, - "learning_rate": 9.239583333333334e-06, - "loss": 0.0121, - "step": 5270 - }, - { - "epoch": 53.99808245445829, - "grad_norm": 0.2920975983142853, - "learning_rate": 9.21875e-06, - "loss": 0.0124, - "step": 5280 - }, - { - "epoch": 54.10035155001598, - "grad_norm": 0.31395605206489563, - "learning_rate": 9.197916666666667e-06, - "loss": 0.0116, - "step": 5290 - }, - { - "epoch": 54.20262064557367, - "grad_norm": 0.3078666627407074, - "learning_rate": 9.177083333333334e-06, - "loss": 0.0117, - "step": 5300 - }, - { - "epoch": 54.20262064557367, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3260598182678223, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7898, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.201, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.025, - "step": 5300 - }, - { - "epoch": 54.304889741131355, - "grad_norm": 0.2747853994369507, - "learning_rate": 9.15625e-06, - "loss": 0.0113, - "step": 5310 - }, - { - "epoch": 54.407158836689035, - "grad_norm": 0.3068675398826599, - "learning_rate": 9.135416666666667e-06, - "loss": 0.0115, - "step": 5320 - }, - { - "epoch": 54.50942793224672, - "grad_norm": 0.29503345489501953, - "learning_rate": 9.114583333333334e-06, - "loss": 0.0116, - "step": 5330 - }, - { - "epoch": 54.61169702780441, - "grad_norm": 0.27636656165122986, - "learning_rate": 9.09375e-06, - "loss": 0.0114, - "step": 5340 - }, - { - "epoch": 54.7139661233621, - "grad_norm": 0.313203901052475, - "learning_rate": 9.072916666666668e-06, - "loss": 0.012, - "step": 5350 - }, - { - "epoch": 54.816235218919786, - "grad_norm": 0.30984997749328613, - "learning_rate": 9.052083333333334e-06, - "loss": 0.012, - "step": 5360 - }, - { - "epoch": 54.91850431447747, - "grad_norm": 0.27004456520080566, - "learning_rate": 9.031250000000001e-06, - "loss": 0.0118, - "step": 5370 - }, - { - "epoch": 55.020773410035154, - "grad_norm": 0.278777152299881, - "learning_rate": 9.010416666666668e-06, - "loss": 0.0117, - "step": 5380 - }, - { - "epoch": 55.12304250559284, - "grad_norm": 0.3070172965526581, - "learning_rate": 8.989583333333334e-06, - "loss": 0.011, - "step": 5390 - }, - { - "epoch": 55.22531160115053, - "grad_norm": 0.2756708562374115, - "learning_rate": 8.968750000000001e-06, - "loss": 0.011, - "step": 5400 - }, - { - "epoch": 55.22531160115053, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.33424711227417, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7849, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.224, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.028, - "step": 5400 - }, - { - "epoch": 55.32758069670821, - "grad_norm": 0.28067439794540405, - "learning_rate": 8.947916666666668e-06, - "loss": 0.011, - "step": 5410 - }, - { - "epoch": 55.4298497922659, - "grad_norm": 0.2702157199382782, - "learning_rate": 8.927083333333335e-06, - "loss": 0.0112, - "step": 5420 - }, - { - "epoch": 55.532118887823586, - "grad_norm": 0.30056026577949524, - "learning_rate": 8.906250000000001e-06, - "loss": 0.0115, - "step": 5430 - }, - { - "epoch": 55.634387983381274, - "grad_norm": 0.29118791222572327, - "learning_rate": 8.885416666666668e-06, - "loss": 0.0113, - "step": 5440 - }, - { - "epoch": 55.73665707893896, - "grad_norm": 0.3045700490474701, - "learning_rate": 8.864583333333335e-06, - "loss": 0.0115, - "step": 5450 - }, - { - "epoch": 55.83892617449664, - "grad_norm": 0.28140169382095337, - "learning_rate": 8.843750000000002e-06, - "loss": 0.0113, - "step": 5460 - }, - { - "epoch": 55.94119527005433, - "grad_norm": 0.2764737010002136, - "learning_rate": 8.822916666666668e-06, - "loss": 0.0118, - "step": 5470 - }, - { - "epoch": 56.04346436561202, - "grad_norm": 0.28735798597335815, - "learning_rate": 8.802083333333335e-06, - "loss": 0.0114, - "step": 5480 - }, - { - "epoch": 56.145733461169705, - "grad_norm": 0.28290146589279175, - "learning_rate": 8.781250000000002e-06, - "loss": 0.0107, - "step": 5490 - }, - { - "epoch": 56.248002556727386, - "grad_norm": 0.2917637526988983, - "learning_rate": 8.760416666666668e-06, - "loss": 0.0108, - "step": 5500 - }, - { - "epoch": 56.248002556727386, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3639163970947266, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7163, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.542, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.068, - "step": 5500 - }, - { - "epoch": 56.35027165228507, - "grad_norm": 0.31157800555229187, - "learning_rate": 8.739583333333333e-06, - "loss": 0.0109, - "step": 5510 - }, - { - "epoch": 56.45254074784276, - "grad_norm": 0.26494109630584717, - "learning_rate": 8.71875e-06, - "loss": 0.0108, - "step": 5520 - }, - { - "epoch": 56.55480984340045, - "grad_norm": 0.2839730381965637, - "learning_rate": 8.697916666666667e-06, - "loss": 0.011, - "step": 5530 - }, - { - "epoch": 56.657078938958136, - "grad_norm": 0.27168846130371094, - "learning_rate": 8.677083333333334e-06, - "loss": 0.0111, - "step": 5540 - }, - { - "epoch": 56.75934803451582, - "grad_norm": 0.2758902311325073, - "learning_rate": 8.65625e-06, - "loss": 0.0109, - "step": 5550 - }, - { - "epoch": 56.861617130073505, - "grad_norm": 0.29986515641212463, - "learning_rate": 8.635416666666667e-06, - "loss": 0.0113, - "step": 5560 - }, - { - "epoch": 56.96388622563119, - "grad_norm": 0.2895634174346924, - "learning_rate": 8.614583333333334e-06, - "loss": 0.0112, - "step": 5570 - }, - { - "epoch": 57.06615532118888, - "grad_norm": 0.26848530769348145, - "learning_rate": 8.59375e-06, - "loss": 0.0105, - "step": 5580 - }, - { - "epoch": 57.16842441674657, - "grad_norm": 0.2801918685436249, - "learning_rate": 8.572916666666667e-06, - "loss": 0.0106, - "step": 5590 - }, - { - "epoch": 57.27069351230425, - "grad_norm": 0.2772628366947174, - "learning_rate": 8.552083333333334e-06, - "loss": 0.0105, - "step": 5600 - }, - { - "epoch": 57.27069351230425, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3490209579467773, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7134, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.556, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.069, - "step": 5600 - }, - { - "epoch": 57.372962607861936, - "grad_norm": 0.26178139448165894, - "learning_rate": 8.53125e-06, - "loss": 0.0103, - "step": 5610 - }, - { - "epoch": 57.475231703419624, - "grad_norm": 0.27127423882484436, - "learning_rate": 8.510416666666667e-06, - "loss": 0.0105, - "step": 5620 - }, - { - "epoch": 57.57750079897731, - "grad_norm": 0.2728956639766693, - "learning_rate": 8.489583333333334e-06, - "loss": 0.0104, - "step": 5630 - }, - { - "epoch": 57.67976989453499, - "grad_norm": 0.26934826374053955, - "learning_rate": 8.468750000000001e-06, - "loss": 0.0107, - "step": 5640 - }, - { - "epoch": 57.78203899009268, - "grad_norm": 0.30065199732780457, - "learning_rate": 8.447916666666668e-06, - "loss": 0.0108, - "step": 5650 - }, - { - "epoch": 57.88430808565037, - "grad_norm": 0.29894405603408813, - "learning_rate": 8.427083333333334e-06, - "loss": 0.0104, - "step": 5660 - }, - { - "epoch": 57.986577181208055, - "grad_norm": 0.26421451568603516, - "learning_rate": 8.406250000000001e-06, - "loss": 0.0105, - "step": 5670 - }, - { - "epoch": 58.08884627676574, - "grad_norm": 0.27239277958869934, - "learning_rate": 8.385416666666668e-06, - "loss": 0.0099, - "step": 5680 - }, - { - "epoch": 58.191115372323424, - "grad_norm": 0.27293631434440613, - "learning_rate": 8.364583333333334e-06, - "loss": 0.01, - "step": 5690 - }, - { - "epoch": 58.29338446788111, - "grad_norm": 0.251788467168808, - "learning_rate": 8.343750000000001e-06, - "loss": 0.0103, - "step": 5700 - }, - { - "epoch": 58.29338446788111, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.360133647918701, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7658, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.312, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.039, - "step": 5700 - }, - { - "epoch": 58.3956535634388, - "grad_norm": 0.2742460072040558, - "learning_rate": 8.322916666666668e-06, - "loss": 0.01, - "step": 5710 - }, - { - "epoch": 58.49792265899649, - "grad_norm": 0.2849868834018707, - "learning_rate": 8.302083333333335e-06, - "loss": 0.0101, - "step": 5720 - }, - { - "epoch": 58.60019175455417, - "grad_norm": 0.2823048233985901, - "learning_rate": 8.281250000000001e-06, - "loss": 0.0101, - "step": 5730 - }, - { - "epoch": 58.702460850111855, - "grad_norm": 0.2816413342952728, - "learning_rate": 8.260416666666668e-06, - "loss": 0.0102, - "step": 5740 - }, - { - "epoch": 58.80472994566954, - "grad_norm": 0.3086257576942444, - "learning_rate": 8.239583333333335e-06, - "loss": 0.0102, - "step": 5750 - }, - { - "epoch": 58.90699904122723, - "grad_norm": 0.31176137924194336, - "learning_rate": 8.218750000000002e-06, - "loss": 0.0102, - "step": 5760 - }, - { - "epoch": 59.00926813678492, - "grad_norm": 0.30020809173583984, - "learning_rate": 8.197916666666668e-06, - "loss": 0.0103, - "step": 5770 - }, - { - "epoch": 59.1115372323426, - "grad_norm": 0.27167460322380066, - "learning_rate": 8.177083333333335e-06, - "loss": 0.0096, - "step": 5780 - }, - { - "epoch": 59.213806327900286, - "grad_norm": 0.29027628898620605, - "learning_rate": 8.156250000000002e-06, - "loss": 0.0099, - "step": 5790 - }, - { - "epoch": 59.316075423457974, - "grad_norm": 0.2502238154411316, - "learning_rate": 8.135416666666668e-06, - "loss": 0.0097, - "step": 5800 - }, - { - "epoch": 59.316075423457974, - "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3818840980529785, - "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7922, - "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.19, - "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.024, - "step": 5800 - } - ], - "logging_steps": 10, - "max_steps": 9700, - "num_input_tokens_seen": 0, - "num_train_epochs": 100, - "save_steps": 200, - "total_flos": 7.772644061346693e+18, - "train_batch_size": 16, - "trial_name": null, - "trial_params": null -}