{ "best_metric": 2.3818840980529785, "best_model_checkpoint": "/ssd1/geonmin.kim/shortened-llm/outputs/phi2_1.8b_alpaca_enzh_fullparam/checkpoint-5800", "epoch": 59.316075423457974, "eval_steps": 100, "global_step": 5800, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.010226909555768616, "grad_norm": 3.242844581604004, "learning_rate": 2.0000000000000002e-07, "loss": 2.0435, "step": 1 }, { "epoch": 0.10226909555768617, "grad_norm": 2.2696735858917236, "learning_rate": 2.0000000000000003e-06, "loss": 2.0397, "step": 10 }, { "epoch": 0.20453819111537233, "grad_norm": 1.283516764640808, "learning_rate": 4.000000000000001e-06, "loss": 1.7851, "step": 20 }, { "epoch": 0.3068072866730585, "grad_norm": 1.1238912343978882, "learning_rate": 6e-06, "loss": 1.6526, "step": 30 }, { "epoch": 0.40907638223074466, "grad_norm": 1.0196110010147095, "learning_rate": 8.000000000000001e-06, "loss": 1.5391, "step": 40 }, { "epoch": 0.5113454777884308, "grad_norm": 0.9775912761688232, "learning_rate": 1e-05, "loss": 1.4779, "step": 50 }, { "epoch": 0.613614573346117, "grad_norm": 0.9060685038566589, "learning_rate": 1.2e-05, "loss": 1.4218, "step": 60 }, { "epoch": 0.7158836689038032, "grad_norm": 0.9413829445838928, "learning_rate": 1.4e-05, "loss": 1.3978, "step": 70 }, { "epoch": 0.8181527644614893, "grad_norm": 1.007612705230713, "learning_rate": 1.6000000000000003e-05, "loss": 1.3779, "step": 80 }, { "epoch": 0.9204218600191755, "grad_norm": 0.9302048087120056, "learning_rate": 1.8e-05, "loss": 1.3387, "step": 90 }, { "epoch": 1.0226909555768615, "grad_norm": 0.82750403881073, "learning_rate": 2e-05, "loss": 1.3056, "step": 100 }, { "epoch": 1.0226909555768615, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.201456069946289, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7448, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.41, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.051, "step": 100 }, { "epoch": 1.124960051134548, "grad_norm": 0.8751218318939209, "learning_rate": 1.9979166666666667e-05, "loss": 1.2376, "step": 110 }, { "epoch": 1.2272291466922338, "grad_norm": 0.8453890681266785, "learning_rate": 1.9958333333333335e-05, "loss": 1.2063, "step": 120 }, { "epoch": 1.3294982422499202, "grad_norm": 0.8156936764717102, "learning_rate": 1.99375e-05, "loss": 1.2022, "step": 130 }, { "epoch": 1.4317673378076063, "grad_norm": 0.8538714051246643, "learning_rate": 1.991666666666667e-05, "loss": 1.2052, "step": 140 }, { "epoch": 1.5340364333652925, "grad_norm": 0.8687230944633484, "learning_rate": 1.9895833333333334e-05, "loss": 1.1913, "step": 150 }, { "epoch": 1.6363055289229784, "grad_norm": 0.826032280921936, "learning_rate": 1.9875000000000002e-05, "loss": 1.1828, "step": 160 }, { "epoch": 1.7385746244806648, "grad_norm": 0.8092572093009949, "learning_rate": 1.9854166666666667e-05, "loss": 1.1746, "step": 170 }, { "epoch": 1.840843720038351, "grad_norm": 0.825042188167572, "learning_rate": 1.9833333333333335e-05, "loss": 1.169, "step": 180 }, { "epoch": 1.943112815596037, "grad_norm": 0.8175085186958313, "learning_rate": 1.98125e-05, "loss": 1.166, "step": 190 }, { "epoch": 2.045381911153723, "grad_norm": 0.8639950752258301, "learning_rate": 1.979166666666667e-05, "loss": 1.1158, "step": 200 }, { "epoch": 2.045381911153723, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1383644342422485, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7181, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.534, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.067, "step": 200 }, { "epoch": 2.1476510067114094, "grad_norm": 0.7958649396896362, "learning_rate": 1.9770833333333334e-05, "loss": 1.0339, "step": 210 }, { "epoch": 2.249920102269096, "grad_norm": 0.8010162711143494, "learning_rate": 1.9750000000000002e-05, "loss": 1.0426, "step": 220 }, { "epoch": 2.3521891978267817, "grad_norm": 0.7885578274726868, "learning_rate": 1.9729166666666667e-05, "loss": 1.0432, "step": 230 }, { "epoch": 2.4544582933844676, "grad_norm": 0.8269402980804443, "learning_rate": 1.9708333333333336e-05, "loss": 1.0379, "step": 240 }, { "epoch": 2.556727388942154, "grad_norm": 0.8139219880104065, "learning_rate": 1.96875e-05, "loss": 1.0305, "step": 250 }, { "epoch": 2.6589964844998404, "grad_norm": 0.8240389227867126, "learning_rate": 1.9666666666666666e-05, "loss": 1.032, "step": 260 }, { "epoch": 2.7612655800575263, "grad_norm": 0.7902543544769287, "learning_rate": 1.9645833333333334e-05, "loss": 1.0203, "step": 270 }, { "epoch": 2.8635346756152127, "grad_norm": 0.7832587361335754, "learning_rate": 1.9625e-05, "loss": 1.0266, "step": 280 }, { "epoch": 2.9658037711728986, "grad_norm": 0.8064606189727783, "learning_rate": 1.9604166666666668e-05, "loss": 1.0292, "step": 290 }, { "epoch": 3.068072866730585, "grad_norm": 0.8083460927009583, "learning_rate": 1.9583333333333333e-05, "loss": 0.9461, "step": 300 }, { "epoch": 3.068072866730585, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1164577007293701, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7799, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.247, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.031, "step": 300 }, { "epoch": 3.170341962288271, "grad_norm": 0.7945894598960876, "learning_rate": 1.95625e-05, "loss": 0.9071, "step": 310 }, { "epoch": 3.2726110578459573, "grad_norm": 0.8095683455467224, "learning_rate": 1.9541666666666666e-05, "loss": 0.9081, "step": 320 }, { "epoch": 3.3748801534036432, "grad_norm": 0.779909074306488, "learning_rate": 1.9520833333333335e-05, "loss": 0.9171, "step": 330 }, { "epoch": 3.4771492489613296, "grad_norm": 0.8173850774765015, "learning_rate": 1.95e-05, "loss": 0.9085, "step": 340 }, { "epoch": 3.5794183445190155, "grad_norm": 0.8204581141471863, "learning_rate": 1.9479166666666668e-05, "loss": 0.9073, "step": 350 }, { "epoch": 3.681687440076702, "grad_norm": 0.7985939383506775, "learning_rate": 1.9458333333333333e-05, "loss": 0.9181, "step": 360 }, { "epoch": 3.783956535634388, "grad_norm": 0.8140767812728882, "learning_rate": 1.94375e-05, "loss": 0.9193, "step": 370 }, { "epoch": 3.886225631192074, "grad_norm": 0.8102839589118958, "learning_rate": 1.9416666666666667e-05, "loss": 0.9157, "step": 380 }, { "epoch": 3.98849472674976, "grad_norm": 0.7886559963226318, "learning_rate": 1.9395833333333335e-05, "loss": 0.9191, "step": 390 }, { "epoch": 4.090763822307446, "grad_norm": 0.7990887761116028, "learning_rate": 1.9375e-05, "loss": 0.8203, "step": 400 }, { "epoch": 4.090763822307446, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1279124021530151, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7222, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.515, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.064, "step": 400 }, { "epoch": 4.193032917865133, "grad_norm": 0.8134937286376953, "learning_rate": 1.935416666666667e-05, "loss": 0.81, "step": 410 }, { "epoch": 4.295302013422819, "grad_norm": 0.8099086284637451, "learning_rate": 1.9333333333333333e-05, "loss": 0.8, "step": 420 }, { "epoch": 4.397571108980505, "grad_norm": 0.7907654047012329, "learning_rate": 1.9312500000000002e-05, "loss": 0.8025, "step": 430 }, { "epoch": 4.499840204538192, "grad_norm": 0.835167407989502, "learning_rate": 1.9291666666666667e-05, "loss": 0.804, "step": 440 }, { "epoch": 4.6021093000958775, "grad_norm": 0.8259727358818054, "learning_rate": 1.9270833333333335e-05, "loss": 0.8069, "step": 450 }, { "epoch": 4.704378395653563, "grad_norm": 0.8095329999923706, "learning_rate": 1.925e-05, "loss": 0.8099, "step": 460 }, { "epoch": 4.806647491211249, "grad_norm": 0.8239871859550476, "learning_rate": 1.922916666666667e-05, "loss": 0.8178, "step": 470 }, { "epoch": 4.908916586768935, "grad_norm": 0.8325999975204468, "learning_rate": 1.9208333333333337e-05, "loss": 0.8077, "step": 480 }, { "epoch": 5.011185682326622, "grad_norm": 0.813887357711792, "learning_rate": 1.9187500000000002e-05, "loss": 0.8061, "step": 490 }, { "epoch": 5.113454777884308, "grad_norm": 0.8288919925689697, "learning_rate": 1.916666666666667e-05, "loss": 0.7063, "step": 500 }, { "epoch": 5.113454777884308, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.1451531648635864, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7807, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.243, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.03, "step": 500 }, { "epoch": 5.215723873441994, "grad_norm": 0.8298630118370056, "learning_rate": 1.9145833333333336e-05, "loss": 0.7049, "step": 510 }, { "epoch": 5.317992968999681, "grad_norm": 0.8246304988861084, "learning_rate": 1.9125000000000004e-05, "loss": 0.7104, "step": 520 }, { "epoch": 5.420262064557367, "grad_norm": 0.7879628539085388, "learning_rate": 1.910416666666667e-05, "loss": 0.7147, "step": 530 }, { "epoch": 5.522531160115053, "grad_norm": 0.8137240409851074, "learning_rate": 1.9083333333333338e-05, "loss": 0.7079, "step": 540 }, { "epoch": 5.624800255672739, "grad_norm": 0.8201763033866882, "learning_rate": 1.9062500000000003e-05, "loss": 0.7162, "step": 550 }, { "epoch": 5.727069351230425, "grad_norm": 0.8055077791213989, "learning_rate": 1.9041666666666668e-05, "loss": 0.7241, "step": 560 }, { "epoch": 5.829338446788111, "grad_norm": 0.8504828810691833, "learning_rate": 1.9020833333333336e-05, "loss": 0.7241, "step": 570 }, { "epoch": 5.931607542345797, "grad_norm": 0.8135042190551758, "learning_rate": 1.9e-05, "loss": 0.7276, "step": 580 }, { "epoch": 6.033876637903483, "grad_norm": 0.907433271408081, "learning_rate": 1.897916666666667e-05, "loss": 0.6957, "step": 590 }, { "epoch": 6.13614573346117, "grad_norm": 0.8348938822746277, "learning_rate": 1.8958333333333334e-05, "loss": 0.6172, "step": 600 }, { "epoch": 6.13614573346117, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.171920657157898, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7497, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.387, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.048, "step": 600 }, { "epoch": 6.238414829018856, "grad_norm": 0.8306870460510254, "learning_rate": 1.8937500000000003e-05, "loss": 0.6255, "step": 610 }, { "epoch": 6.340683924576542, "grad_norm": 0.8185003399848938, "learning_rate": 1.8916666666666668e-05, "loss": 0.6218, "step": 620 }, { "epoch": 6.442953020134228, "grad_norm": 0.8396403789520264, "learning_rate": 1.8895833333333336e-05, "loss": 0.6282, "step": 630 }, { "epoch": 6.545222115691915, "grad_norm": 0.8167343139648438, "learning_rate": 1.8875e-05, "loss": 0.6305, "step": 640 }, { "epoch": 6.6474912112496005, "grad_norm": 0.8276931047439575, "learning_rate": 1.885416666666667e-05, "loss": 0.6378, "step": 650 }, { "epoch": 6.7497603068072864, "grad_norm": 0.8405306339263916, "learning_rate": 1.8833333333333335e-05, "loss": 0.6375, "step": 660 }, { "epoch": 6.852029402364973, "grad_norm": 0.8206018805503845, "learning_rate": 1.8812500000000003e-05, "loss": 0.6393, "step": 670 }, { "epoch": 6.954298497922659, "grad_norm": 0.8008025884628296, "learning_rate": 1.8791666666666668e-05, "loss": 0.6448, "step": 680 }, { "epoch": 7.056567593480345, "grad_norm": 0.8231362104415894, "learning_rate": 1.8770833333333337e-05, "loss": 0.5929, "step": 690 }, { "epoch": 7.158836689038031, "grad_norm": 0.8469756841659546, "learning_rate": 1.8750000000000002e-05, "loss": 0.5392, "step": 700 }, { "epoch": 7.158836689038031, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.2256364822387695, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8506, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.92, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.99, "step": 700 }, { "epoch": 7.261105784595718, "grad_norm": 0.8122310042381287, "learning_rate": 1.8729166666666667e-05, "loss": 0.54, "step": 710 }, { "epoch": 7.363374880153404, "grad_norm": 0.8621034026145935, "learning_rate": 1.8708333333333335e-05, "loss": 0.5534, "step": 720 }, { "epoch": 7.46564397571109, "grad_norm": 0.8529708981513977, "learning_rate": 1.86875e-05, "loss": 0.5516, "step": 730 }, { "epoch": 7.567913071268776, "grad_norm": 0.8307532072067261, "learning_rate": 1.866666666666667e-05, "loss": 0.5531, "step": 740 }, { "epoch": 7.6701821668264625, "grad_norm": 0.8403590321540833, "learning_rate": 1.8645833333333334e-05, "loss": 0.5644, "step": 750 }, { "epoch": 7.772451262384148, "grad_norm": 0.8332897424697876, "learning_rate": 1.8625000000000002e-05, "loss": 0.5607, "step": 760 }, { "epoch": 7.874720357941834, "grad_norm": 0.866201639175415, "learning_rate": 1.8604166666666667e-05, "loss": 0.5658, "step": 770 }, { "epoch": 7.97698945349952, "grad_norm": 0.8362734913825989, "learning_rate": 1.8583333333333336e-05, "loss": 0.566, "step": 780 }, { "epoch": 8.079258549057206, "grad_norm": 0.8327781558036804, "learning_rate": 1.85625e-05, "loss": 0.4963, "step": 790 }, { "epoch": 8.181527644614892, "grad_norm": 0.827268123626709, "learning_rate": 1.854166666666667e-05, "loss": 0.4752, "step": 800 }, { "epoch": 8.181527644614892, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.2643427848815918, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.893, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.726, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.966, "step": 800 }, { "epoch": 8.28379674017258, "grad_norm": 0.7902690172195435, "learning_rate": 1.8520833333333334e-05, "loss": 0.4843, "step": 810 }, { "epoch": 8.386065835730266, "grad_norm": 0.8375966548919678, "learning_rate": 1.8500000000000002e-05, "loss": 0.4783, "step": 820 }, { "epoch": 8.488334931287952, "grad_norm": 0.8248066306114197, "learning_rate": 1.8479166666666667e-05, "loss": 0.4785, "step": 830 }, { "epoch": 8.590604026845638, "grad_norm": 0.8647124171257019, "learning_rate": 1.8458333333333336e-05, "loss": 0.4874, "step": 840 }, { "epoch": 8.692873122403324, "grad_norm": 0.8471198678016663, "learning_rate": 1.84375e-05, "loss": 0.4884, "step": 850 }, { "epoch": 8.79514221796101, "grad_norm": 0.8394653797149658, "learning_rate": 1.8416666666666666e-05, "loss": 0.497, "step": 860 }, { "epoch": 8.897411313518695, "grad_norm": 0.8621006011962891, "learning_rate": 1.8395833333333334e-05, "loss": 0.4923, "step": 870 }, { "epoch": 8.999680409076383, "grad_norm": 0.830193817615509, "learning_rate": 1.8375e-05, "loss": 0.4917, "step": 880 }, { "epoch": 9.101949504634069, "grad_norm": 0.8418065309524536, "learning_rate": 1.8354166666666668e-05, "loss": 0.4104, "step": 890 }, { "epoch": 9.204218600191755, "grad_norm": 0.8672240376472473, "learning_rate": 1.8333333333333333e-05, "loss": 0.4165, "step": 900 }, { "epoch": 9.204218600191755, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.2995867729187012, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8906, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.737, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.967, "step": 900 }, { "epoch": 9.306487695749441, "grad_norm": 0.8118753433227539, "learning_rate": 1.83125e-05, "loss": 0.4181, "step": 910 }, { "epoch": 9.408756791307127, "grad_norm": 0.8521497845649719, "learning_rate": 1.8291666666666666e-05, "loss": 0.4202, "step": 920 }, { "epoch": 9.511025886864813, "grad_norm": 0.8396993279457092, "learning_rate": 1.8270833333333335e-05, "loss": 0.4247, "step": 930 }, { "epoch": 9.613294982422499, "grad_norm": 0.8380371332168579, "learning_rate": 1.825e-05, "loss": 0.427, "step": 940 }, { "epoch": 9.715564077980185, "grad_norm": 0.819240927696228, "learning_rate": 1.8229166666666668e-05, "loss": 0.4271, "step": 950 }, { "epoch": 9.817833173537872, "grad_norm": 0.8467490077018738, "learning_rate": 1.8208333333333333e-05, "loss": 0.4288, "step": 960 }, { "epoch": 9.920102269095558, "grad_norm": 0.8679558634757996, "learning_rate": 1.81875e-05, "loss": 0.4283, "step": 970 }, { "epoch": 10.022371364653244, "grad_norm": 0.8336887359619141, "learning_rate": 1.8166666666666667e-05, "loss": 0.4244, "step": 980 }, { "epoch": 10.12464046021093, "grad_norm": 0.8790706396102905, "learning_rate": 1.8145833333333335e-05, "loss": 0.3627, "step": 990 }, { "epoch": 10.226909555768616, "grad_norm": 0.8260719180107117, "learning_rate": 1.8125e-05, "loss": 0.3622, "step": 1000 }, { "epoch": 10.226909555768616, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.3518736362457275, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.891, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.735, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.967, "step": 1000 }, { "epoch": 10.329178651326302, "grad_norm": 0.8354102969169617, "learning_rate": 1.810416666666667e-05, "loss": 0.3654, "step": 1010 }, { "epoch": 10.431447746883988, "grad_norm": 0.833742618560791, "learning_rate": 1.8083333333333334e-05, "loss": 0.3675, "step": 1020 }, { "epoch": 10.533716842441674, "grad_norm": 0.814929723739624, "learning_rate": 1.8062500000000002e-05, "loss": 0.368, "step": 1030 }, { "epoch": 10.635985937999362, "grad_norm": 0.8500260710716248, "learning_rate": 1.8041666666666667e-05, "loss": 0.3704, "step": 1040 }, { "epoch": 10.738255033557047, "grad_norm": 0.8280666470527649, "learning_rate": 1.8020833333333335e-05, "loss": 0.3722, "step": 1050 }, { "epoch": 10.840524129114733, "grad_norm": 0.871941089630127, "learning_rate": 1.8e-05, "loss": 0.3768, "step": 1060 }, { "epoch": 10.94279322467242, "grad_norm": 0.8509662747383118, "learning_rate": 1.797916666666667e-05, "loss": 0.3829, "step": 1070 }, { "epoch": 11.045062320230105, "grad_norm": 0.8621676564216614, "learning_rate": 1.7958333333333334e-05, "loss": 0.3512, "step": 1080 }, { "epoch": 11.147331415787791, "grad_norm": 0.8415457010269165, "learning_rate": 1.7937500000000002e-05, "loss": 0.3092, "step": 1090 }, { "epoch": 11.249600511345477, "grad_norm": 0.8087013363838196, "learning_rate": 1.7916666666666667e-05, "loss": 0.3158, "step": 1100 }, { "epoch": 11.249600511345477, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.4024713039398193, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8756, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.805, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.976, "step": 1100 }, { "epoch": 11.351869606903165, "grad_norm": 0.8239210844039917, "learning_rate": 1.7895833333333336e-05, "loss": 0.3189, "step": 1110 }, { "epoch": 11.45413870246085, "grad_norm": 0.8607499003410339, "learning_rate": 1.7875e-05, "loss": 0.3222, "step": 1120 }, { "epoch": 11.556407798018537, "grad_norm": 0.8538540601730347, "learning_rate": 1.785416666666667e-05, "loss": 0.3207, "step": 1130 }, { "epoch": 11.658676893576223, "grad_norm": 0.8388919234275818, "learning_rate": 1.7833333333333334e-05, "loss": 0.3308, "step": 1140 }, { "epoch": 11.760945989133909, "grad_norm": 0.8096144795417786, "learning_rate": 1.7812500000000003e-05, "loss": 0.3256, "step": 1150 }, { "epoch": 11.863215084691594, "grad_norm": 0.8199524879455566, "learning_rate": 1.7791666666666668e-05, "loss": 0.3272, "step": 1160 }, { "epoch": 11.96548418024928, "grad_norm": 0.8503059148788452, "learning_rate": 1.7770833333333336e-05, "loss": 0.3313, "step": 1170 }, { "epoch": 12.067753275806966, "grad_norm": 0.8376160860061646, "learning_rate": 1.775e-05, "loss": 0.2907, "step": 1180 }, { "epoch": 12.170022371364654, "grad_norm": 5.976499557495117, "learning_rate": 1.772916666666667e-05, "loss": 0.2719, "step": 1190 }, { "epoch": 12.27229146692234, "grad_norm": 0.8097366094589233, "learning_rate": 1.7708333333333335e-05, "loss": 0.2782, "step": 1200 }, { "epoch": 12.27229146692234, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.4481781721115112, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.9199, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.603, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.95, "step": 1200 }, { "epoch": 12.374560562480026, "grad_norm": 0.7708520293235779, "learning_rate": 1.7687500000000003e-05, "loss": 0.2749, "step": 1210 }, { "epoch": 12.476829658037712, "grad_norm": 0.8598915338516235, "learning_rate": 1.7666666666666668e-05, "loss": 0.2784, "step": 1220 }, { "epoch": 12.579098753595398, "grad_norm": 0.8157161474227905, "learning_rate": 1.7645833333333336e-05, "loss": 0.2794, "step": 1230 }, { "epoch": 12.681367849153084, "grad_norm": 0.8496010899543762, "learning_rate": 1.7625e-05, "loss": 0.2825, "step": 1240 }, { "epoch": 12.78363694471077, "grad_norm": 0.815390944480896, "learning_rate": 1.760416666666667e-05, "loss": 0.2848, "step": 1250 }, { "epoch": 12.885906040268456, "grad_norm": 0.8204165697097778, "learning_rate": 1.7583333333333335e-05, "loss": 0.2883, "step": 1260 }, { "epoch": 12.988175135826143, "grad_norm": 0.8374896049499512, "learning_rate": 1.7562500000000003e-05, "loss": 0.2909, "step": 1270 }, { "epoch": 13.09044423138383, "grad_norm": 0.7642439007759094, "learning_rate": 1.754166666666667e-05, "loss": 0.2406, "step": 1280 }, { "epoch": 13.192713326941515, "grad_norm": 0.7860681414604187, "learning_rate": 1.7520833333333337e-05, "loss": 0.2358, "step": 1290 }, { "epoch": 13.294982422499201, "grad_norm": 0.8119321465492249, "learning_rate": 1.7500000000000002e-05, "loss": 0.2357, "step": 1300 }, { "epoch": 13.294982422499201, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.4959287643432617, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8771, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.799, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.975, "step": 1300 }, { "epoch": 13.397251518056887, "grad_norm": 0.7898840308189392, "learning_rate": 1.7479166666666667e-05, "loss": 0.2409, "step": 1310 }, { "epoch": 13.499520613614573, "grad_norm": 0.8298600912094116, "learning_rate": 1.7458333333333335e-05, "loss": 0.2399, "step": 1320 }, { "epoch": 13.601789709172259, "grad_norm": 0.7994723916053772, "learning_rate": 1.74375e-05, "loss": 0.2468, "step": 1330 }, { "epoch": 13.704058804729947, "grad_norm": 0.822475790977478, "learning_rate": 1.741666666666667e-05, "loss": 0.2481, "step": 1340 }, { "epoch": 13.806327900287632, "grad_norm": 0.8012453317642212, "learning_rate": 1.7395833333333334e-05, "loss": 0.2473, "step": 1350 }, { "epoch": 13.908596995845318, "grad_norm": 0.8046063780784607, "learning_rate": 1.7375000000000002e-05, "loss": 0.2532, "step": 1360 }, { "epoch": 14.010866091403004, "grad_norm": 0.7180681228637695, "learning_rate": 1.7354166666666667e-05, "loss": 0.2451, "step": 1370 }, { "epoch": 14.11313518696069, "grad_norm": 0.7648767828941345, "learning_rate": 1.7333333333333336e-05, "loss": 0.2026, "step": 1380 }, { "epoch": 14.215404282518376, "grad_norm": 0.782311201095581, "learning_rate": 1.73125e-05, "loss": 0.2062, "step": 1390 }, { "epoch": 14.317673378076062, "grad_norm": 0.7766838669776917, "learning_rate": 1.729166666666667e-05, "loss": 0.2079, "step": 1400 }, { "epoch": 14.317673378076062, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.535245418548584, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8983, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.701, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.963, "step": 1400 }, { "epoch": 14.419942473633748, "grad_norm": 0.7595117688179016, "learning_rate": 1.7270833333333334e-05, "loss": 0.2097, "step": 1410 }, { "epoch": 14.522211569191436, "grad_norm": 0.7640486359596252, "learning_rate": 1.7250000000000003e-05, "loss": 0.2107, "step": 1420 }, { "epoch": 14.624480664749122, "grad_norm": 0.8100217580795288, "learning_rate": 1.7229166666666668e-05, "loss": 0.2132, "step": 1430 }, { "epoch": 14.726749760306808, "grad_norm": 0.7824357748031616, "learning_rate": 1.7208333333333336e-05, "loss": 0.2124, "step": 1440 }, { "epoch": 14.829018855864494, "grad_norm": 0.8215783834457397, "learning_rate": 1.71875e-05, "loss": 0.2172, "step": 1450 }, { "epoch": 14.93128795142218, "grad_norm": 0.791244626045227, "learning_rate": 1.7166666666666666e-05, "loss": 0.2183, "step": 1460 }, { "epoch": 15.033557046979865, "grad_norm": 0.7542420625686646, "learning_rate": 1.7145833333333334e-05, "loss": 0.2083, "step": 1470 }, { "epoch": 15.135826142537551, "grad_norm": 0.7965226173400879, "learning_rate": 1.7125e-05, "loss": 0.1773, "step": 1480 }, { "epoch": 15.238095238095237, "grad_norm": 0.764574408531189, "learning_rate": 1.7104166666666668e-05, "loss": 0.1775, "step": 1490 }, { "epoch": 15.340364333652925, "grad_norm": 0.768020749092102, "learning_rate": 1.7083333333333333e-05, "loss": 0.181, "step": 1500 }, { "epoch": 15.340364333652925, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.57305908203125, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7396, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.434, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.054, "step": 1500 }, { "epoch": 15.44263342921061, "grad_norm": 0.7567213773727417, "learning_rate": 1.70625e-05, "loss": 0.181, "step": 1510 }, { "epoch": 15.544902524768297, "grad_norm": 0.7426446080207825, "learning_rate": 1.7041666666666666e-05, "loss": 0.1824, "step": 1520 }, { "epoch": 15.647171620325983, "grad_norm": 0.750170111656189, "learning_rate": 1.7020833333333335e-05, "loss": 0.1855, "step": 1530 }, { "epoch": 15.749440715883669, "grad_norm": 0.7680428624153137, "learning_rate": 1.7e-05, "loss": 0.1883, "step": 1540 }, { "epoch": 15.851709811441355, "grad_norm": 0.8081603646278381, "learning_rate": 1.6979166666666668e-05, "loss": 0.1892, "step": 1550 }, { "epoch": 15.95397890699904, "grad_norm": 0.7882938385009766, "learning_rate": 1.6958333333333333e-05, "loss": 0.1912, "step": 1560 }, { "epoch": 16.05624800255673, "grad_norm": 0.7245915532112122, "learning_rate": 1.6937500000000002e-05, "loss": 0.1717, "step": 1570 }, { "epoch": 16.158517098114412, "grad_norm": 0.7572883367538452, "learning_rate": 1.6916666666666667e-05, "loss": 0.1577, "step": 1580 }, { "epoch": 16.2607861936721, "grad_norm": 0.7576362490653992, "learning_rate": 1.6895833333333335e-05, "loss": 0.157, "step": 1590 }, { "epoch": 16.363055289229784, "grad_norm": 0.7182960510253906, "learning_rate": 1.6875e-05, "loss": 0.1555, "step": 1600 }, { "epoch": 16.363055289229784, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.6099534034729004, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8255, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.036, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.005, "step": 1600 }, { "epoch": 16.465324384787472, "grad_norm": 0.7401660084724426, "learning_rate": 1.685416666666667e-05, "loss": 0.1573, "step": 1610 }, { "epoch": 16.56759348034516, "grad_norm": 0.7596891522407532, "learning_rate": 1.6833333333333334e-05, "loss": 0.1589, "step": 1620 }, { "epoch": 16.669862575902844, "grad_norm": 0.767320990562439, "learning_rate": 1.6812500000000002e-05, "loss": 0.1613, "step": 1630 }, { "epoch": 16.77213167146053, "grad_norm": 0.7579568028450012, "learning_rate": 1.6791666666666667e-05, "loss": 0.1624, "step": 1640 }, { "epoch": 16.874400767018216, "grad_norm": 0.7488529682159424, "learning_rate": 1.6770833333333336e-05, "loss": 0.1646, "step": 1650 }, { "epoch": 16.976669862575903, "grad_norm": 0.7528676390647888, "learning_rate": 1.675e-05, "loss": 0.1686, "step": 1660 }, { "epoch": 17.078938958133588, "grad_norm": 0.6930709481239319, "learning_rate": 1.672916666666667e-05, "loss": 0.1429, "step": 1670 }, { "epoch": 17.181208053691275, "grad_norm": 0.7119250297546387, "learning_rate": 1.6708333333333334e-05, "loss": 0.1355, "step": 1680 }, { "epoch": 17.283477149248963, "grad_norm": 0.6774548292160034, "learning_rate": 1.6687500000000002e-05, "loss": 0.1348, "step": 1690 }, { "epoch": 17.385746244806647, "grad_norm": 0.7192063927650452, "learning_rate": 1.6666666666666667e-05, "loss": 0.1368, "step": 1700 }, { "epoch": 17.385746244806647, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.6659198999404907, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8832, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.771, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.971, "step": 1700 }, { "epoch": 17.488015340364335, "grad_norm": 0.6981900930404663, "learning_rate": 1.6645833333333336e-05, "loss": 0.1379, "step": 1710 }, { "epoch": 17.59028443592202, "grad_norm": 0.7018482685089111, "learning_rate": 1.6625e-05, "loss": 0.1413, "step": 1720 }, { "epoch": 17.692553531479707, "grad_norm": 0.7282826900482178, "learning_rate": 1.660416666666667e-05, "loss": 0.1437, "step": 1730 }, { "epoch": 17.79482262703739, "grad_norm": 0.751104474067688, "learning_rate": 1.6583333333333334e-05, "loss": 0.142, "step": 1740 }, { "epoch": 17.89709172259508, "grad_norm": 0.7333133816719055, "learning_rate": 1.6562500000000003e-05, "loss": 0.1452, "step": 1750 }, { "epoch": 17.999360818152766, "grad_norm": 0.7490417957305908, "learning_rate": 1.6541666666666668e-05, "loss": 0.1445, "step": 1760 }, { "epoch": 18.10162991371045, "grad_norm": 0.7038053274154663, "learning_rate": 1.6520833333333336e-05, "loss": 0.1196, "step": 1770 }, { "epoch": 18.203899009268138, "grad_norm": 0.6416111588478088, "learning_rate": 1.65e-05, "loss": 0.1189, "step": 1780 }, { "epoch": 18.306168104825822, "grad_norm": 0.6799057126045227, "learning_rate": 1.647916666666667e-05, "loss": 0.1187, "step": 1790 }, { "epoch": 18.40843720038351, "grad_norm": 0.70688396692276, "learning_rate": 1.6458333333333335e-05, "loss": 0.1195, "step": 1800 }, { "epoch": 18.40843720038351, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.701189398765564, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8449, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.947, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.993, "step": 1800 }, { "epoch": 18.510706295941194, "grad_norm": 0.6970006823539734, "learning_rate": 1.6437500000000003e-05, "loss": 0.1223, "step": 1810 }, { "epoch": 18.612975391498882, "grad_norm": 0.7127917408943176, "learning_rate": 1.6416666666666668e-05, "loss": 0.1228, "step": 1820 }, { "epoch": 18.715244487056566, "grad_norm": 0.6875537633895874, "learning_rate": 1.6395833333333337e-05, "loss": 0.1234, "step": 1830 }, { "epoch": 18.817513582614254, "grad_norm": 0.7249884009361267, "learning_rate": 1.6375e-05, "loss": 0.126, "step": 1840 }, { "epoch": 18.91978267817194, "grad_norm": 0.7007323503494263, "learning_rate": 1.635416666666667e-05, "loss": 0.1243, "step": 1850 }, { "epoch": 19.022051773729626, "grad_norm": 0.6626154184341431, "learning_rate": 1.6333333333333335e-05, "loss": 0.1233, "step": 1860 }, { "epoch": 19.124320869287313, "grad_norm": 0.6599249243736267, "learning_rate": 1.6312500000000003e-05, "loss": 0.1034, "step": 1870 }, { "epoch": 19.226589964844997, "grad_norm": 0.6787338256835938, "learning_rate": 1.629166666666667e-05, "loss": 0.1031, "step": 1880 }, { "epoch": 19.328859060402685, "grad_norm": 0.6463894248008728, "learning_rate": 1.6270833333333337e-05, "loss": 0.106, "step": 1890 }, { "epoch": 19.43112815596037, "grad_norm": 0.6649991869926453, "learning_rate": 1.6250000000000002e-05, "loss": 0.1063, "step": 1900 }, { "epoch": 19.43112815596037, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.7399890422821045, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8883, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.747, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.968, "step": 1900 }, { "epoch": 19.533397251518057, "grad_norm": 0.6542336940765381, "learning_rate": 1.6229166666666667e-05, "loss": 0.1059, "step": 1910 }, { "epoch": 19.635666347075745, "grad_norm": 0.6741412281990051, "learning_rate": 1.6208333333333335e-05, "loss": 0.1088, "step": 1920 }, { "epoch": 19.73793544263343, "grad_norm": 0.6898075938224792, "learning_rate": 1.61875e-05, "loss": 0.1087, "step": 1930 }, { "epoch": 19.840204538191117, "grad_norm": 0.6611754298210144, "learning_rate": 1.616666666666667e-05, "loss": 0.1098, "step": 1940 }, { "epoch": 19.9424736337488, "grad_norm": 0.6986993551254272, "learning_rate": 1.6145833333333334e-05, "loss": 0.1114, "step": 1950 }, { "epoch": 20.04474272930649, "grad_norm": 0.6218557953834534, "learning_rate": 1.6125000000000002e-05, "loss": 0.1028, "step": 1960 }, { "epoch": 20.147011824864173, "grad_norm": 0.6389386653900146, "learning_rate": 1.6104166666666667e-05, "loss": 0.0922, "step": 1970 }, { "epoch": 20.24928092042186, "grad_norm": 0.6259350776672363, "learning_rate": 1.6083333333333336e-05, "loss": 0.0924, "step": 1980 }, { "epoch": 20.351550015979548, "grad_norm": 0.6429017782211304, "learning_rate": 1.60625e-05, "loss": 0.0935, "step": 1990 }, { "epoch": 20.453819111537232, "grad_norm": 0.6409590244293213, "learning_rate": 1.604166666666667e-05, "loss": 0.0925, "step": 2000 }, { "epoch": 20.453819111537232, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.770869493484497, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8821, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.776, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.972, "step": 2000 }, { "epoch": 20.55608820709492, "grad_norm": 0.6365486979484558, "learning_rate": 1.6020833333333334e-05, "loss": 0.0945, "step": 2010 }, { "epoch": 20.658357302652604, "grad_norm": 0.6637431979179382, "learning_rate": 1.6000000000000003e-05, "loss": 0.0963, "step": 2020 }, { "epoch": 20.76062639821029, "grad_norm": 0.6256012916564941, "learning_rate": 1.5979166666666668e-05, "loss": 0.0981, "step": 2030 }, { "epoch": 20.862895493767976, "grad_norm": 0.6379542946815491, "learning_rate": 1.5958333333333336e-05, "loss": 0.0976, "step": 2040 }, { "epoch": 20.965164589325664, "grad_norm": 0.6680212020874023, "learning_rate": 1.59375e-05, "loss": 0.0982, "step": 2050 }, { "epoch": 21.067433684883348, "grad_norm": 0.6050080060958862, "learning_rate": 1.5916666666666666e-05, "loss": 0.0876, "step": 2060 }, { "epoch": 21.169702780441035, "grad_norm": 0.6315256953239441, "learning_rate": 1.5895833333333335e-05, "loss": 0.0824, "step": 2070 }, { "epoch": 21.271971875998723, "grad_norm": 0.6160369515419006, "learning_rate": 1.5875e-05, "loss": 0.0817, "step": 2080 }, { "epoch": 21.374240971556407, "grad_norm": 0.5967450737953186, "learning_rate": 1.5854166666666668e-05, "loss": 0.0828, "step": 2090 }, { "epoch": 21.476510067114095, "grad_norm": 0.6033092737197876, "learning_rate": 1.5833333333333333e-05, "loss": 0.0832, "step": 2100 }, { "epoch": 21.476510067114095, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.8033993244171143, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8632, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.862, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.983, "step": 2100 }, { "epoch": 21.57877916267178, "grad_norm": 0.6197848916053772, "learning_rate": 1.58125e-05, "loss": 0.0852, "step": 2110 }, { "epoch": 21.681048258229467, "grad_norm": 0.6332718133926392, "learning_rate": 1.5791666666666667e-05, "loss": 0.0853, "step": 2120 }, { "epoch": 21.78331735378715, "grad_norm": 0.6289674043655396, "learning_rate": 1.5770833333333335e-05, "loss": 0.0861, "step": 2130 }, { "epoch": 21.88558644934484, "grad_norm": 0.6466374397277832, "learning_rate": 1.575e-05, "loss": 0.0873, "step": 2140 }, { "epoch": 21.987855544902526, "grad_norm": 2.8282721042633057, "learning_rate": 1.5733333333333334e-05, "loss": 0.0888, "step": 2150 }, { "epoch": 22.09012464046021, "grad_norm": 0.5769690275192261, "learning_rate": 1.5712500000000002e-05, "loss": 0.0748, "step": 2160 }, { "epoch": 22.192393736017898, "grad_norm": 0.5819457173347473, "learning_rate": 1.5691666666666667e-05, "loss": 0.0737, "step": 2170 }, { "epoch": 22.294662831575582, "grad_norm": 0.6134530305862427, "learning_rate": 1.5670833333333336e-05, "loss": 0.0739, "step": 2180 }, { "epoch": 22.39693192713327, "grad_norm": 0.6075708866119385, "learning_rate": 1.565e-05, "loss": 0.0742, "step": 2190 }, { "epoch": 22.499201022690954, "grad_norm": 0.5924075245857239, "learning_rate": 1.562916666666667e-05, "loss": 0.0751, "step": 2200 }, { "epoch": 22.499201022690954, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.829688549041748, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8883, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.747, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.968, "step": 2200 }, { "epoch": 22.601470118248642, "grad_norm": 0.6079824566841125, "learning_rate": 1.5608333333333334e-05, "loss": 0.0762, "step": 2210 }, { "epoch": 22.70373921380633, "grad_norm": 0.6125743389129639, "learning_rate": 1.5587500000000003e-05, "loss": 0.0764, "step": 2220 }, { "epoch": 22.806008309364014, "grad_norm": 0.5798956751823425, "learning_rate": 1.5566666666666668e-05, "loss": 0.0764, "step": 2230 }, { "epoch": 22.9082774049217, "grad_norm": 0.6115660667419434, "learning_rate": 1.5545833333333336e-05, "loss": 0.0782, "step": 2240 }, { "epoch": 23.010546500479386, "grad_norm": 0.5898561477661133, "learning_rate": 1.5525e-05, "loss": 0.0785, "step": 2250 }, { "epoch": 23.112815596037073, "grad_norm": 0.5677723288536072, "learning_rate": 1.550416666666667e-05, "loss": 0.066, "step": 2260 }, { "epoch": 23.215084691594758, "grad_norm": 0.5659050345420837, "learning_rate": 1.5483333333333335e-05, "loss": 0.0659, "step": 2270 }, { "epoch": 23.317353787152445, "grad_norm": 0.5879548192024231, "learning_rate": 1.54625e-05, "loss": 0.0671, "step": 2280 }, { "epoch": 23.41962288271013, "grad_norm": 0.5900695323944092, "learning_rate": 1.5441666666666668e-05, "loss": 0.0668, "step": 2290 }, { "epoch": 23.521891978267817, "grad_norm": 0.5663672685623169, "learning_rate": 1.5420833333333333e-05, "loss": 0.0676, "step": 2300 }, { "epoch": 23.521891978267817, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.8641977310180664, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8705, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.829, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.979, "step": 2300 }, { "epoch": 23.624161073825505, "grad_norm": 0.6123341917991638, "learning_rate": 1.54e-05, "loss": 0.0686, "step": 2310 }, { "epoch": 23.72643016938319, "grad_norm": 0.5857561826705933, "learning_rate": 1.5379166666666667e-05, "loss": 0.0688, "step": 2320 }, { "epoch": 23.828699264940877, "grad_norm": 0.6261973977088928, "learning_rate": 1.5358333333333335e-05, "loss": 0.0694, "step": 2330 }, { "epoch": 23.93096836049856, "grad_norm": 0.5833300948143005, "learning_rate": 1.53375e-05, "loss": 0.0706, "step": 2340 }, { "epoch": 24.03323745605625, "grad_norm": 0.5474048256874084, "learning_rate": 1.531666666666667e-05, "loss": 0.0676, "step": 2350 }, { "epoch": 24.135506551613933, "grad_norm": 0.5484219193458557, "learning_rate": 1.5295833333333334e-05, "loss": 0.06, "step": 2360 }, { "epoch": 24.23777564717162, "grad_norm": 0.5514199137687683, "learning_rate": 1.5275000000000002e-05, "loss": 0.0608, "step": 2370 }, { "epoch": 24.340044742729308, "grad_norm": 0.5593263506889343, "learning_rate": 1.5254166666666667e-05, "loss": 0.0607, "step": 2380 }, { "epoch": 24.442313838286992, "grad_norm": 0.5621811151504517, "learning_rate": 1.5233333333333335e-05, "loss": 0.0612, "step": 2390 }, { "epoch": 24.54458293384468, "grad_norm": 0.5351541042327881, "learning_rate": 1.52125e-05, "loss": 0.0612, "step": 2400 }, { "epoch": 24.54458293384468, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.8911913633346558, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8888, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.745, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.968, "step": 2400 }, { "epoch": 24.646852029402364, "grad_norm": 0.6047305464744568, "learning_rate": 1.5191666666666669e-05, "loss": 0.0628, "step": 2410 }, { "epoch": 24.749121124960052, "grad_norm": 0.5636226534843445, "learning_rate": 1.5170833333333334e-05, "loss": 0.0624, "step": 2420 }, { "epoch": 24.851390220517736, "grad_norm": 0.5656375288963318, "learning_rate": 1.515e-05, "loss": 0.0635, "step": 2430 }, { "epoch": 24.953659316075424, "grad_norm": 0.5728236436843872, "learning_rate": 1.5129166666666667e-05, "loss": 0.0642, "step": 2440 }, { "epoch": 25.05592841163311, "grad_norm": 0.5383201837539673, "learning_rate": 1.5108333333333334e-05, "loss": 0.0585, "step": 2450 }, { "epoch": 25.158197507190796, "grad_norm": 0.5213799476623535, "learning_rate": 1.50875e-05, "loss": 0.0539, "step": 2460 }, { "epoch": 25.260466602748483, "grad_norm": 0.5278561115264893, "learning_rate": 1.5066666666666668e-05, "loss": 0.0557, "step": 2470 }, { "epoch": 25.362735698306167, "grad_norm": 0.5517110824584961, "learning_rate": 1.5045833333333334e-05, "loss": 0.054, "step": 2480 }, { "epoch": 25.465004793863855, "grad_norm": 0.5266678333282471, "learning_rate": 1.5025000000000001e-05, "loss": 0.055, "step": 2490 }, { "epoch": 25.56727388942154, "grad_norm": 0.5280548334121704, "learning_rate": 1.5004166666666668e-05, "loss": 0.056, "step": 2500 }, { "epoch": 25.56727388942154, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.9079244136810303, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8724, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.82, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.978, "step": 2500 }, { "epoch": 25.669542984979227, "grad_norm": 0.543380081653595, "learning_rate": 1.4983333333333334e-05, "loss": 0.0567, "step": 2510 }, { "epoch": 25.77181208053691, "grad_norm": 0.5638805031776428, "learning_rate": 1.4962500000000003e-05, "loss": 0.0576, "step": 2520 }, { "epoch": 25.8740811760946, "grad_norm": 0.5791529417037964, "learning_rate": 1.4941666666666668e-05, "loss": 0.0585, "step": 2530 }, { "epoch": 25.976350271652286, "grad_norm": 0.5550780892372131, "learning_rate": 1.4920833333333336e-05, "loss": 0.0583, "step": 2540 }, { "epoch": 26.07861936720997, "grad_norm": 0.5031822323799133, "learning_rate": 1.4900000000000001e-05, "loss": 0.0521, "step": 2550 }, { "epoch": 26.18088846276766, "grad_norm": 0.5042924880981445, "learning_rate": 1.487916666666667e-05, "loss": 0.0495, "step": 2560 }, { "epoch": 26.283157558325343, "grad_norm": 0.5483749508857727, "learning_rate": 1.4858333333333335e-05, "loss": 0.0504, "step": 2570 }, { "epoch": 26.38542665388303, "grad_norm": 0.508752703666687, "learning_rate": 1.48375e-05, "loss": 0.0509, "step": 2580 }, { "epoch": 26.487695749440714, "grad_norm": 0.4737156331539154, "learning_rate": 1.4816666666666668e-05, "loss": 0.0507, "step": 2590 }, { "epoch": 26.589964844998402, "grad_norm": 0.5402019023895264, "learning_rate": 1.4795833333333333e-05, "loss": 0.0513, "step": 2600 }, { "epoch": 26.589964844998402, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.9507412910461426, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8703, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.83, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.979, "step": 2600 }, { "epoch": 26.69223394055609, "grad_norm": 0.5165488123893738, "learning_rate": 1.4775000000000002e-05, "loss": 0.0519, "step": 2610 }, { "epoch": 26.794503036113774, "grad_norm": 0.5218144059181213, "learning_rate": 1.4754166666666667e-05, "loss": 0.0537, "step": 2620 }, { "epoch": 26.89677213167146, "grad_norm": 0.5503118634223938, "learning_rate": 1.4733333333333335e-05, "loss": 0.053, "step": 2630 }, { "epoch": 26.999041227229146, "grad_norm": 0.5653933882713318, "learning_rate": 1.47125e-05, "loss": 0.0533, "step": 2640 }, { "epoch": 27.101310322786833, "grad_norm": 0.5077876448631287, "learning_rate": 1.4691666666666669e-05, "loss": 0.0473, "step": 2650 }, { "epoch": 27.203579418344518, "grad_norm": 0.5198752880096436, "learning_rate": 1.4670833333333334e-05, "loss": 0.0472, "step": 2660 }, { "epoch": 27.305848513902205, "grad_norm": 0.49260616302490234, "learning_rate": 1.4650000000000002e-05, "loss": 0.047, "step": 2670 }, { "epoch": 27.408117609459893, "grad_norm": 0.5261008739471436, "learning_rate": 1.4629166666666667e-05, "loss": 0.0468, "step": 2680 }, { "epoch": 27.510386705017577, "grad_norm": 0.5464609265327454, "learning_rate": 1.4608333333333335e-05, "loss": 0.0476, "step": 2690 }, { "epoch": 27.612655800575265, "grad_norm": 0.526250422000885, "learning_rate": 1.45875e-05, "loss": 0.0488, "step": 2700 }, { "epoch": 27.612655800575265, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.963590145111084, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8709, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.827, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.978, "step": 2700 }, { "epoch": 27.71492489613295, "grad_norm": 0.49739035964012146, "learning_rate": 1.4566666666666669e-05, "loss": 0.0478, "step": 2710 }, { "epoch": 27.817193991690637, "grad_norm": 0.5307214260101318, "learning_rate": 1.4545833333333334e-05, "loss": 0.048, "step": 2720 }, { "epoch": 27.91946308724832, "grad_norm": 0.5277805328369141, "learning_rate": 1.4525e-05, "loss": 0.0489, "step": 2730 }, { "epoch": 28.02173218280601, "grad_norm": 0.490689218044281, "learning_rate": 1.4504166666666667e-05, "loss": 0.0488, "step": 2740 }, { "epoch": 28.124001278363693, "grad_norm": 0.5121079683303833, "learning_rate": 1.4483333333333334e-05, "loss": 0.0426, "step": 2750 }, { "epoch": 28.22627037392138, "grad_norm": 0.5197770595550537, "learning_rate": 1.4462500000000001e-05, "loss": 0.0433, "step": 2760 }, { "epoch": 28.328539469479068, "grad_norm": 0.5023459196090698, "learning_rate": 1.4441666666666668e-05, "loss": 0.0428, "step": 2770 }, { "epoch": 28.430808565036752, "grad_norm": 0.47794410586357117, "learning_rate": 1.4420833333333334e-05, "loss": 0.0431, "step": 2780 }, { "epoch": 28.53307766059444, "grad_norm": 0.4780057370662689, "learning_rate": 1.4400000000000001e-05, "loss": 0.0439, "step": 2790 }, { "epoch": 28.635346756152124, "grad_norm": 0.5096541047096252, "learning_rate": 1.4379166666666668e-05, "loss": 0.0443, "step": 2800 }, { "epoch": 28.635346756152124, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 1.9908784627914429, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8556, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.897, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.987, "step": 2800 }, { "epoch": 28.737615851709812, "grad_norm": 0.49939730763435364, "learning_rate": 1.4358333333333334e-05, "loss": 0.045, "step": 2810 }, { "epoch": 28.839884947267496, "grad_norm": 0.5106296539306641, "learning_rate": 1.4337500000000001e-05, "loss": 0.045, "step": 2820 }, { "epoch": 28.942154042825184, "grad_norm": 0.4849531054496765, "learning_rate": 1.4316666666666668e-05, "loss": 0.0451, "step": 2830 }, { "epoch": 29.04442313838287, "grad_norm": 0.46746712923049927, "learning_rate": 1.4295833333333335e-05, "loss": 0.0429, "step": 2840 }, { "epoch": 29.146692233940556, "grad_norm": 0.4314974546432495, "learning_rate": 1.4275000000000001e-05, "loss": 0.0397, "step": 2850 }, { "epoch": 29.248961329498243, "grad_norm": 0.4507407546043396, "learning_rate": 1.4254166666666668e-05, "loss": 0.0394, "step": 2860 }, { "epoch": 29.351230425055927, "grad_norm": 0.4565179646015167, "learning_rate": 1.4233333333333335e-05, "loss": 0.0402, "step": 2870 }, { "epoch": 29.453499520613615, "grad_norm": 0.4820208251476288, "learning_rate": 1.42125e-05, "loss": 0.0403, "step": 2880 }, { "epoch": 29.5557686161713, "grad_norm": 0.47718533873558044, "learning_rate": 1.4191666666666668e-05, "loss": 0.0404, "step": 2890 }, { "epoch": 29.658037711728987, "grad_norm": 0.5102068185806274, "learning_rate": 1.4170833333333333e-05, "loss": 0.0418, "step": 2900 }, { "epoch": 29.658037711728987, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.013613700866699, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8531, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.909, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.989, "step": 2900 }, { "epoch": 29.760306807286675, "grad_norm": 0.49850982427597046, "learning_rate": 1.4150000000000002e-05, "loss": 0.041, "step": 2910 }, { "epoch": 29.86257590284436, "grad_norm": 0.5028281211853027, "learning_rate": 1.4129166666666667e-05, "loss": 0.0419, "step": 2920 }, { "epoch": 29.964844998402047, "grad_norm": 0.4753844141960144, "learning_rate": 1.4108333333333335e-05, "loss": 0.0421, "step": 2930 }, { "epoch": 30.06711409395973, "grad_norm": 0.45387542247772217, "learning_rate": 1.40875e-05, "loss": 0.0381, "step": 2940 }, { "epoch": 30.16938318951742, "grad_norm": 0.4576801359653473, "learning_rate": 1.4066666666666669e-05, "loss": 0.0365, "step": 2950 }, { "epoch": 30.271652285075103, "grad_norm": 1.5916332006454468, "learning_rate": 1.4045833333333334e-05, "loss": 0.0371, "step": 2960 }, { "epoch": 30.37392138063279, "grad_norm": 28.43450164794922, "learning_rate": 1.4029166666666668e-05, "loss": 0.0403, "step": 2970 }, { "epoch": 30.476190476190474, "grad_norm": 23.308544158935547, "learning_rate": 1.4010416666666669e-05, "loss": 0.3001, "step": 2980 }, { "epoch": 30.578459571748162, "grad_norm": 0.9334068298339844, "learning_rate": 1.3989583333333334e-05, "loss": 0.2114, "step": 2990 }, { "epoch": 30.68072866730585, "grad_norm": 0.5433509945869446, "learning_rate": 1.3968750000000002e-05, "loss": 0.0448, "step": 3000 }, { "epoch": 30.68072866730585, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.014845609664917, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8553, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.899, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.987, "step": 3000 }, { "epoch": 30.782997762863534, "grad_norm": 0.5084338188171387, "learning_rate": 1.3947916666666667e-05, "loss": 0.042, "step": 3010 }, { "epoch": 30.88526685842122, "grad_norm": 0.4836946725845337, "learning_rate": 1.3927083333333336e-05, "loss": 0.0396, "step": 3020 }, { "epoch": 30.987535953978906, "grad_norm": 0.4835638701915741, "learning_rate": 1.3906250000000001e-05, "loss": 0.0397, "step": 3030 }, { "epoch": 31.089805049536594, "grad_norm": 0.4484618008136749, "learning_rate": 1.3885416666666666e-05, "loss": 0.0359, "step": 3040 }, { "epoch": 31.192074145094278, "grad_norm": 0.4576917886734009, "learning_rate": 1.3864583333333334e-05, "loss": 0.0352, "step": 3050 }, { "epoch": 31.294343240651965, "grad_norm": 0.4618643820285797, "learning_rate": 1.3843750000000001e-05, "loss": 0.0351, "step": 3060 }, { "epoch": 31.396612336209653, "grad_norm": 0.4532334804534912, "learning_rate": 1.3822916666666668e-05, "loss": 0.0353, "step": 3070 }, { "epoch": 31.498881431767337, "grad_norm": 0.43523749709129333, "learning_rate": 1.3802083333333335e-05, "loss": 0.0357, "step": 3080 }, { "epoch": 31.601150527325025, "grad_norm": 0.4621034860610962, "learning_rate": 1.3781250000000001e-05, "loss": 0.036, "step": 3090 }, { "epoch": 31.70341962288271, "grad_norm": 0.4407444894313812, "learning_rate": 1.3760416666666668e-05, "loss": 0.0365, "step": 3100 }, { "epoch": 31.70341962288271, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.0446879863739014, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8565, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.893, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.987, "step": 3100 }, { "epoch": 31.805688718440397, "grad_norm": 0.46584275364875793, "learning_rate": 1.3739583333333335e-05, "loss": 0.0367, "step": 3110 }, { "epoch": 31.90795781399808, "grad_norm": 0.4602925181388855, "learning_rate": 1.3718750000000001e-05, "loss": 0.0368, "step": 3120 }, { "epoch": 32.01022690955577, "grad_norm": 0.4346022605895996, "learning_rate": 1.3697916666666668e-05, "loss": 0.0398, "step": 3130 }, { "epoch": 32.11249600511346, "grad_norm": 0.42625728249549866, "learning_rate": 1.3677083333333335e-05, "loss": 0.0327, "step": 3140 }, { "epoch": 32.214765100671144, "grad_norm": 0.4581526219844818, "learning_rate": 1.3656250000000002e-05, "loss": 0.0325, "step": 3150 }, { "epoch": 32.317034196228825, "grad_norm": 0.41788893938064575, "learning_rate": 1.3635416666666668e-05, "loss": 0.0323, "step": 3160 }, { "epoch": 32.41930329178651, "grad_norm": 0.39856305718421936, "learning_rate": 1.3614583333333335e-05, "loss": 0.0331, "step": 3170 }, { "epoch": 32.5215723873442, "grad_norm": 0.4417785704135895, "learning_rate": 1.3593750000000002e-05, "loss": 0.0336, "step": 3180 }, { "epoch": 32.62384148290189, "grad_norm": 0.45321398973464966, "learning_rate": 1.3572916666666667e-05, "loss": 0.0339, "step": 3190 }, { "epoch": 32.72611057845957, "grad_norm": 0.43776148557662964, "learning_rate": 1.3552083333333335e-05, "loss": 0.0337, "step": 3200 }, { "epoch": 32.72611057845957, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.0593974590301514, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8377, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.98, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.997, "step": 3200 }, { "epoch": 32.828379674017256, "grad_norm": 0.4577961564064026, "learning_rate": 1.353125e-05, "loss": 0.0342, "step": 3210 }, { "epoch": 32.930648769574944, "grad_norm": 0.43651047348976135, "learning_rate": 1.3510416666666669e-05, "loss": 0.0352, "step": 3220 }, { "epoch": 33.03291786513263, "grad_norm": 0.4300122559070587, "learning_rate": 1.3489583333333334e-05, "loss": 0.0331, "step": 3230 }, { "epoch": 33.13518696069032, "grad_norm": 0.43906641006469727, "learning_rate": 1.3468750000000002e-05, "loss": 0.0304, "step": 3240 }, { "epoch": 33.237456056248, "grad_norm": 0.40852677822113037, "learning_rate": 1.3447916666666667e-05, "loss": 0.0304, "step": 3250 }, { "epoch": 33.33972515180569, "grad_norm": 0.42338284850120544, "learning_rate": 1.3427083333333336e-05, "loss": 0.0307, "step": 3260 }, { "epoch": 33.441994247363375, "grad_norm": 0.4441679120063782, "learning_rate": 1.340625e-05, "loss": 0.031, "step": 3270 }, { "epoch": 33.54426334292106, "grad_norm": 0.431755393743515, "learning_rate": 1.3385416666666669e-05, "loss": 0.031, "step": 3280 }, { "epoch": 33.646532438478744, "grad_norm": 0.42783161997795105, "learning_rate": 1.3364583333333334e-05, "loss": 0.0316, "step": 3290 }, { "epoch": 33.74880153403643, "grad_norm": 0.4471502900123596, "learning_rate": 1.3343750000000002e-05, "loss": 0.0318, "step": 3300 }, { "epoch": 33.74880153403643, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.0793018341064453, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.9135, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.632, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.954, "step": 3300 }, { "epoch": 33.85107062959412, "grad_norm": 0.42543718218803406, "learning_rate": 1.3322916666666668e-05, "loss": 0.0317, "step": 3310 }, { "epoch": 33.95333972515181, "grad_norm": 0.4455837905406952, "learning_rate": 1.3302083333333336e-05, "loss": 0.0321, "step": 3320 }, { "epoch": 34.055608820709494, "grad_norm": 0.451027512550354, "learning_rate": 1.3281250000000001e-05, "loss": 0.0297, "step": 3330 }, { "epoch": 34.157877916267175, "grad_norm": 0.43288454413414, "learning_rate": 1.3260416666666666e-05, "loss": 0.0285, "step": 3340 }, { "epoch": 34.26014701182486, "grad_norm": 0.41506046056747437, "learning_rate": 1.3239583333333334e-05, "loss": 0.0285, "step": 3350 }, { "epoch": 34.36241610738255, "grad_norm": 0.4198153614997864, "learning_rate": 1.321875e-05, "loss": 0.0287, "step": 3360 }, { "epoch": 34.46468520294024, "grad_norm": 0.4576322138309479, "learning_rate": 1.3197916666666668e-05, "loss": 0.0293, "step": 3370 }, { "epoch": 34.566954298497926, "grad_norm": 0.4268178343772888, "learning_rate": 1.3177083333333333e-05, "loss": 0.0297, "step": 3380 }, { "epoch": 34.66922339405561, "grad_norm": 0.4406622648239136, "learning_rate": 1.3156250000000001e-05, "loss": 0.0298, "step": 3390 }, { "epoch": 34.771492489613294, "grad_norm": 0.44503384828567505, "learning_rate": 1.3135416666666666e-05, "loss": 0.0302, "step": 3400 }, { "epoch": 34.771492489613294, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1051576137542725, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7904, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.198, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.025, "step": 3400 }, { "epoch": 34.87376158517098, "grad_norm": 0.4410327672958374, "learning_rate": 1.3114583333333335e-05, "loss": 0.0308, "step": 3410 }, { "epoch": 34.97603068072867, "grad_norm": 0.4527595639228821, "learning_rate": 1.309375e-05, "loss": 0.0307, "step": 3420 }, { "epoch": 35.07829977628635, "grad_norm": 0.3983278274536133, "learning_rate": 1.3072916666666668e-05, "loss": 0.0275, "step": 3430 }, { "epoch": 35.18056887184404, "grad_norm": 0.405274361371994, "learning_rate": 1.3052083333333335e-05, "loss": 0.027, "step": 3440 }, { "epoch": 35.282837967401726, "grad_norm": 0.4225080907344818, "learning_rate": 1.3031250000000002e-05, "loss": 0.0276, "step": 3450 }, { "epoch": 35.38510706295941, "grad_norm": 0.3988070487976074, "learning_rate": 1.3010416666666668e-05, "loss": 0.0271, "step": 3460 }, { "epoch": 35.4873761585171, "grad_norm": 0.613850474357605, "learning_rate": 1.2989583333333335e-05, "loss": 0.0271, "step": 3470 }, { "epoch": 35.58964525407478, "grad_norm": 0.42930155992507935, "learning_rate": 1.2968750000000002e-05, "loss": 0.0269, "step": 3480 }, { "epoch": 35.69191434963247, "grad_norm": 0.4209059178829193, "learning_rate": 1.2947916666666667e-05, "loss": 0.0281, "step": 3490 }, { "epoch": 35.79418344519016, "grad_norm": 0.4427293837070465, "learning_rate": 1.2927083333333335e-05, "loss": 0.0282, "step": 3500 }, { "epoch": 35.79418344519016, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1160192489624023, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8212, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.056, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.007, "step": 3500 }, { "epoch": 35.896452540747845, "grad_norm": 0.42366185784339905, "learning_rate": 1.290625e-05, "loss": 0.0289, "step": 3510 }, { "epoch": 35.99872163630553, "grad_norm": 0.4120742380619049, "learning_rate": 1.2885416666666669e-05, "loss": 0.0287, "step": 3520 }, { "epoch": 36.10099073186321, "grad_norm": 0.38890495896339417, "learning_rate": 1.2864583333333334e-05, "loss": 0.0258, "step": 3530 }, { "epoch": 36.2032598274209, "grad_norm": 0.4076452851295471, "learning_rate": 1.2843750000000002e-05, "loss": 0.0258, "step": 3540 }, { "epoch": 36.30552892297859, "grad_norm": 0.3809320032596588, "learning_rate": 1.2822916666666667e-05, "loss": 0.0255, "step": 3550 }, { "epoch": 36.407798018536276, "grad_norm": 0.39650237560272217, "learning_rate": 1.2802083333333336e-05, "loss": 0.026, "step": 3560 }, { "epoch": 36.51006711409396, "grad_norm": 0.4009132385253906, "learning_rate": 1.278125e-05, "loss": 0.0265, "step": 3570 }, { "epoch": 36.612336209651644, "grad_norm": 0.42308393120765686, "learning_rate": 1.2760416666666669e-05, "loss": 0.0263, "step": 3580 }, { "epoch": 36.71460530520933, "grad_norm": 0.4044691324234009, "learning_rate": 1.2739583333333334e-05, "loss": 0.0265, "step": 3590 }, { "epoch": 36.81687440076702, "grad_norm": 0.4140937030315399, "learning_rate": 1.2718750000000003e-05, "loss": 0.0266, "step": 3600 }, { "epoch": 36.81687440076702, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.13891339302063, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8236, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.045, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.006, "step": 3600 }, { "epoch": 36.91914349632471, "grad_norm": 0.4160206615924835, "learning_rate": 1.2697916666666668e-05, "loss": 0.0269, "step": 3610 }, { "epoch": 37.02141259188239, "grad_norm": 0.39423078298568726, "learning_rate": 1.2677083333333336e-05, "loss": 0.026, "step": 3620 }, { "epoch": 37.123681687440076, "grad_norm": 0.3859294056892395, "learning_rate": 1.2656250000000001e-05, "loss": 0.0238, "step": 3630 }, { "epoch": 37.225950782997764, "grad_norm": 0.40413331985473633, "learning_rate": 1.2635416666666666e-05, "loss": 0.0244, "step": 3640 }, { "epoch": 37.32821987855545, "grad_norm": 0.39002159237861633, "learning_rate": 1.2614583333333334e-05, "loss": 0.0244, "step": 3650 }, { "epoch": 37.43048897411313, "grad_norm": 0.403145432472229, "learning_rate": 1.259375e-05, "loss": 0.0245, "step": 3660 }, { "epoch": 37.53275806967082, "grad_norm": 0.42878827452659607, "learning_rate": 1.2572916666666668e-05, "loss": 0.0249, "step": 3670 }, { "epoch": 37.63502716522851, "grad_norm": 0.4047834277153015, "learning_rate": 1.2552083333333333e-05, "loss": 0.0246, "step": 3680 }, { "epoch": 37.737296260786195, "grad_norm": 0.4242531955242157, "learning_rate": 1.2531250000000001e-05, "loss": 0.0254, "step": 3690 }, { "epoch": 37.83956535634388, "grad_norm": 0.3883196711540222, "learning_rate": 1.2510416666666666e-05, "loss": 0.025, "step": 3700 }, { "epoch": 37.83956535634388, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.135497570037842, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.828, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.024, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.003, "step": 3700 }, { "epoch": 37.94183445190156, "grad_norm": 0.40664049983024597, "learning_rate": 1.2489583333333335e-05, "loss": 0.0249, "step": 3710 }, { "epoch": 38.04410354745925, "grad_norm": 0.37315633893013, "learning_rate": 1.246875e-05, "loss": 0.0242, "step": 3720 }, { "epoch": 38.14637264301694, "grad_norm": 0.37339115142822266, "learning_rate": 1.2447916666666668e-05, "loss": 0.0229, "step": 3730 }, { "epoch": 38.24864173857463, "grad_norm": 0.38532310724258423, "learning_rate": 1.2427083333333333e-05, "loss": 0.0227, "step": 3740 }, { "epoch": 38.350910834132314, "grad_norm": 0.3730473220348358, "learning_rate": 1.2406250000000002e-05, "loss": 0.0227, "step": 3750 }, { "epoch": 38.453179929689995, "grad_norm": 0.37635692954063416, "learning_rate": 1.2385416666666667e-05, "loss": 0.0234, "step": 3760 }, { "epoch": 38.55544902524768, "grad_norm": 0.38997572660446167, "learning_rate": 1.2364583333333335e-05, "loss": 0.0233, "step": 3770 }, { "epoch": 38.65771812080537, "grad_norm": 0.38781270384788513, "learning_rate": 1.234375e-05, "loss": 0.0237, "step": 3780 }, { "epoch": 38.75998721636306, "grad_norm": 0.42525115609169006, "learning_rate": 1.2322916666666667e-05, "loss": 0.0237, "step": 3790 }, { "epoch": 38.86225631192074, "grad_norm": 0.3896453380584717, "learning_rate": 1.2302083333333335e-05, "loss": 0.0239, "step": 3800 }, { "epoch": 38.86225631192074, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.166919231414795, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8189, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.067, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.008, "step": 3800 }, { "epoch": 38.964525407478426, "grad_norm": 0.3976113498210907, "learning_rate": 1.228125e-05, "loss": 0.0242, "step": 3810 }, { "epoch": 39.066794503036114, "grad_norm": 0.37402892112731934, "learning_rate": 1.2260416666666669e-05, "loss": 0.0228, "step": 3820 }, { "epoch": 39.1690635985938, "grad_norm": 0.4018329679965973, "learning_rate": 1.2239583333333334e-05, "loss": 0.0219, "step": 3830 }, { "epoch": 39.27133269415149, "grad_norm": 0.3905806839466095, "learning_rate": 1.2218750000000002e-05, "loss": 0.0219, "step": 3840 }, { "epoch": 39.37360178970917, "grad_norm": 0.39368128776550293, "learning_rate": 1.2197916666666667e-05, "loss": 0.0223, "step": 3850 }, { "epoch": 39.47587088526686, "grad_norm": 0.40275242924690247, "learning_rate": 1.2177083333333336e-05, "loss": 0.0225, "step": 3860 }, { "epoch": 39.578139980824545, "grad_norm": 0.377655953168869, "learning_rate": 1.215625e-05, "loss": 0.0223, "step": 3870 }, { "epoch": 39.68040907638223, "grad_norm": 0.35910850763320923, "learning_rate": 1.2135416666666669e-05, "loss": 0.0222, "step": 3880 }, { "epoch": 39.782678171939914, "grad_norm": 0.40253758430480957, "learning_rate": 1.2114583333333334e-05, "loss": 0.0229, "step": 3890 }, { "epoch": 39.8849472674976, "grad_norm": 0.4119781255722046, "learning_rate": 1.2093750000000003e-05, "loss": 0.0231, "step": 3900 }, { "epoch": 39.8849472674976, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1834068298339844, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8776, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.796, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.975, "step": 3900 }, { "epoch": 39.98721636305529, "grad_norm": 0.40584734082221985, "learning_rate": 1.2072916666666668e-05, "loss": 0.0229, "step": 3910 }, { "epoch": 40.08948545861298, "grad_norm": 0.3695628345012665, "learning_rate": 1.2052083333333336e-05, "loss": 0.0207, "step": 3920 }, { "epoch": 40.191754554170664, "grad_norm": 0.38678476214408875, "learning_rate": 1.2031250000000001e-05, "loss": 0.0205, "step": 3930 }, { "epoch": 40.294023649728345, "grad_norm": 0.349589079618454, "learning_rate": 1.2010416666666666e-05, "loss": 0.0208, "step": 3940 }, { "epoch": 40.39629274528603, "grad_norm": 0.3602575659751892, "learning_rate": 1.1989583333333335e-05, "loss": 0.0207, "step": 3950 }, { "epoch": 40.49856184084372, "grad_norm": 0.34940171241760254, "learning_rate": 1.196875e-05, "loss": 0.0208, "step": 3960 }, { "epoch": 40.60083093640141, "grad_norm": 0.42674344778060913, "learning_rate": 1.1947916666666668e-05, "loss": 0.0214, "step": 3970 }, { "epoch": 40.703100031959096, "grad_norm": 0.3709782361984253, "learning_rate": 1.1927083333333333e-05, "loss": 0.0215, "step": 3980 }, { "epoch": 40.80536912751678, "grad_norm": 0.37002718448638916, "learning_rate": 1.1906250000000001e-05, "loss": 0.0215, "step": 3990 }, { "epoch": 40.907638223074464, "grad_norm": 0.3808917999267578, "learning_rate": 1.1885416666666666e-05, "loss": 0.0216, "step": 4000 }, { "epoch": 40.907638223074464, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.173332452774048, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8356, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.99, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.999, "step": 4000 }, { "epoch": 41.00990731863215, "grad_norm": 0.5391054153442383, "learning_rate": 1.1864583333333335e-05, "loss": 0.0214, "step": 4010 }, { "epoch": 41.11217641418984, "grad_norm": 0.36861568689346313, "learning_rate": 1.184375e-05, "loss": 0.0203, "step": 4020 }, { "epoch": 41.21444550974752, "grad_norm": 0.3627229630947113, "learning_rate": 1.1822916666666668e-05, "loss": 0.0203, "step": 4030 }, { "epoch": 41.31671460530521, "grad_norm": 0.37450557947158813, "learning_rate": 1.1802083333333333e-05, "loss": 0.02, "step": 4040 }, { "epoch": 41.418983700862896, "grad_norm": 0.3615454435348511, "learning_rate": 1.1781250000000002e-05, "loss": 0.0204, "step": 4050 }, { "epoch": 41.52125279642058, "grad_norm": 0.3542068302631378, "learning_rate": 1.1760416666666667e-05, "loss": 0.0203, "step": 4060 }, { "epoch": 41.62352189197827, "grad_norm": 0.3920552134513855, "learning_rate": 1.1739583333333335e-05, "loss": 0.0204, "step": 4070 }, { "epoch": 41.72579098753595, "grad_norm": 0.3835306763648987, "learning_rate": 1.171875e-05, "loss": 0.0206, "step": 4080 }, { "epoch": 41.82806008309364, "grad_norm": 0.35131242871284485, "learning_rate": 1.1697916666666667e-05, "loss": 0.0208, "step": 4090 }, { "epoch": 41.93032917865133, "grad_norm": 0.3543274402618408, "learning_rate": 1.1677083333333334e-05, "loss": 0.0202, "step": 4100 }, { "epoch": 41.93032917865133, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.1988096237182617, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8148, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.086, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.011, "step": 4100 }, { "epoch": 42.032598274209015, "grad_norm": 0.33442866802215576, "learning_rate": 1.165625e-05, "loss": 0.0198, "step": 4110 }, { "epoch": 42.134867369766695, "grad_norm": 0.3324146866798401, "learning_rate": 1.1635416666666667e-05, "loss": 0.0192, "step": 4120 }, { "epoch": 42.23713646532438, "grad_norm": 0.3776349723339081, "learning_rate": 1.1614583333333334e-05, "loss": 0.0191, "step": 4130 }, { "epoch": 42.33940556088207, "grad_norm": 0.3571579158306122, "learning_rate": 1.159375e-05, "loss": 0.0192, "step": 4140 }, { "epoch": 42.44167465643976, "grad_norm": 0.37327298521995544, "learning_rate": 1.1572916666666667e-05, "loss": 0.0192, "step": 4150 }, { "epoch": 42.543943751997446, "grad_norm": 0.3787640631198883, "learning_rate": 1.1552083333333334e-05, "loss": 0.0196, "step": 4160 }, { "epoch": 42.64621284755513, "grad_norm": 0.36109158396720886, "learning_rate": 1.153125e-05, "loss": 0.0193, "step": 4170 }, { "epoch": 42.748481943112814, "grad_norm": 0.34873998165130615, "learning_rate": 1.151041666666667e-05, "loss": 0.0194, "step": 4180 }, { "epoch": 42.8507510386705, "grad_norm": 0.393928587436676, "learning_rate": 1.1489583333333334e-05, "loss": 0.0196, "step": 4190 }, { "epoch": 42.95302013422819, "grad_norm": 0.3687518537044525, "learning_rate": 1.1468750000000003e-05, "loss": 0.0198, "step": 4200 }, { "epoch": 42.95302013422819, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2081563472747803, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7395, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.434, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.054, "step": 4200 }, { "epoch": 43.05528922978588, "grad_norm": 0.3610316514968872, "learning_rate": 1.1447916666666668e-05, "loss": 0.0187, "step": 4210 }, { "epoch": 43.15755832534356, "grad_norm": 0.3438417911529541, "learning_rate": 1.1427083333333336e-05, "loss": 0.0181, "step": 4220 }, { "epoch": 43.259827420901246, "grad_norm": 0.33607029914855957, "learning_rate": 1.1406250000000001e-05, "loss": 0.018, "step": 4230 }, { "epoch": 43.362096516458934, "grad_norm": 0.3655332326889038, "learning_rate": 1.1385416666666666e-05, "loss": 0.018, "step": 4240 }, { "epoch": 43.46436561201662, "grad_norm": 0.3450184762477875, "learning_rate": 1.1364583333333335e-05, "loss": 0.0182, "step": 4250 }, { "epoch": 43.5666347075743, "grad_norm": 0.34371188282966614, "learning_rate": 1.134375e-05, "loss": 0.0185, "step": 4260 }, { "epoch": 43.66890380313199, "grad_norm": 0.3620694875717163, "learning_rate": 1.1322916666666668e-05, "loss": 0.0188, "step": 4270 }, { "epoch": 43.77117289868968, "grad_norm": 0.34563344717025757, "learning_rate": 1.1302083333333333e-05, "loss": 0.0185, "step": 4280 }, { "epoch": 43.873441994247365, "grad_norm": 0.3630271255970001, "learning_rate": 1.1281250000000001e-05, "loss": 0.0189, "step": 4290 }, { "epoch": 43.97571108980505, "grad_norm": 0.3588694930076599, "learning_rate": 1.1260416666666666e-05, "loss": 0.0192, "step": 4300 }, { "epoch": 43.97571108980505, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.21140456199646, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7918, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.192, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.024, "step": 4300 }, { "epoch": 44.07798018536273, "grad_norm": 0.3162562847137451, "learning_rate": 1.1239583333333335e-05, "loss": 0.0172, "step": 4310 }, { "epoch": 44.18024928092042, "grad_norm": 0.366394966840744, "learning_rate": 1.121875e-05, "loss": 0.017, "step": 4320 }, { "epoch": 44.28251837647811, "grad_norm": 0.35083746910095215, "learning_rate": 1.1197916666666668e-05, "loss": 0.0178, "step": 4330 }, { "epoch": 44.384787472035796, "grad_norm": 0.35160747170448303, "learning_rate": 1.1177083333333333e-05, "loss": 0.0174, "step": 4340 }, { "epoch": 44.48705656759348, "grad_norm": 0.35368478298187256, "learning_rate": 1.1156250000000002e-05, "loss": 0.0175, "step": 4350 }, { "epoch": 44.589325663151165, "grad_norm": 0.3595126271247864, "learning_rate": 1.1135416666666667e-05, "loss": 0.0176, "step": 4360 }, { "epoch": 44.69159475870885, "grad_norm": 0.34251803159713745, "learning_rate": 1.1114583333333335e-05, "loss": 0.0179, "step": 4370 }, { "epoch": 44.79386385426654, "grad_norm": 0.3488711416721344, "learning_rate": 1.109375e-05, "loss": 0.018, "step": 4380 }, { "epoch": 44.89613294982423, "grad_norm": 0.35304006934165955, "learning_rate": 1.1072916666666667e-05, "loss": 0.018, "step": 4390 }, { "epoch": 44.99840204538191, "grad_norm": 0.3306570053100586, "learning_rate": 1.1052083333333334e-05, "loss": 0.018, "step": 4400 }, { "epoch": 44.99840204538191, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.226471424102783, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7401, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.432, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.054, "step": 4400 }, { "epoch": 45.100671140939596, "grad_norm": 0.3414161801338196, "learning_rate": 1.103125e-05, "loss": 0.0163, "step": 4410 }, { "epoch": 45.202940236497284, "grad_norm": 0.34503790736198425, "learning_rate": 1.1010416666666667e-05, "loss": 0.0167, "step": 4420 }, { "epoch": 45.30520933205497, "grad_norm": 0.3578576147556305, "learning_rate": 1.0989583333333334e-05, "loss": 0.0165, "step": 4430 }, { "epoch": 45.40747842761266, "grad_norm": 0.34795865416526794, "learning_rate": 1.096875e-05, "loss": 0.0165, "step": 4440 }, { "epoch": 45.50974752317034, "grad_norm": 0.35079431533813477, "learning_rate": 1.0947916666666667e-05, "loss": 0.0168, "step": 4450 }, { "epoch": 45.61201661872803, "grad_norm": 0.30789047479629517, "learning_rate": 1.0927083333333334e-05, "loss": 0.0164, "step": 4460 }, { "epoch": 45.714285714285715, "grad_norm": 0.370721697807312, "learning_rate": 1.090625e-05, "loss": 0.0174, "step": 4470 }, { "epoch": 45.8165548098434, "grad_norm": 0.34695321321487427, "learning_rate": 1.0885416666666668e-05, "loss": 0.0171, "step": 4480 }, { "epoch": 45.918823905401084, "grad_norm": 0.3586573600769043, "learning_rate": 1.0864583333333334e-05, "loss": 0.0171, "step": 4490 }, { "epoch": 46.02109300095877, "grad_norm": 0.3352271318435669, "learning_rate": 1.0843750000000001e-05, "loss": 0.0172, "step": 4500 }, { "epoch": 46.02109300095877, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2414653301239014, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.694, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.646, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.081, "step": 4500 }, { "epoch": 46.12336209651646, "grad_norm": 0.3106061518192291, "learning_rate": 1.0822916666666668e-05, "loss": 0.0154, "step": 4510 }, { "epoch": 46.22563119207415, "grad_norm": 0.3251485824584961, "learning_rate": 1.0802083333333334e-05, "loss": 0.0159, "step": 4520 }, { "epoch": 46.327900287631834, "grad_norm": 0.3339298367500305, "learning_rate": 1.0781250000000001e-05, "loss": 0.0157, "step": 4530 }, { "epoch": 46.430169383189515, "grad_norm": 0.33735695481300354, "learning_rate": 1.0760416666666666e-05, "loss": 0.0161, "step": 4540 }, { "epoch": 46.5324384787472, "grad_norm": 0.3049178719520569, "learning_rate": 1.0739583333333335e-05, "loss": 0.0162, "step": 4550 }, { "epoch": 46.63470757430489, "grad_norm": 0.3524739742279053, "learning_rate": 1.071875e-05, "loss": 0.016, "step": 4560 }, { "epoch": 46.73697666986258, "grad_norm": 0.3321375250816345, "learning_rate": 1.0697916666666668e-05, "loss": 0.016, "step": 4570 }, { "epoch": 46.83924576542026, "grad_norm": 0.3346993029117584, "learning_rate": 1.0677083333333333e-05, "loss": 0.0165, "step": 4580 }, { "epoch": 46.94151486097795, "grad_norm": 0.3200349807739258, "learning_rate": 1.0656250000000002e-05, "loss": 0.0162, "step": 4590 }, { "epoch": 47.043783956535634, "grad_norm": 0.3281605541706085, "learning_rate": 1.0635416666666667e-05, "loss": 0.0158, "step": 4600 }, { "epoch": 47.043783956535634, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.256087303161621, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.744, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.413, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.052, "step": 4600 }, { "epoch": 47.14605305209332, "grad_norm": 0.35677453875541687, "learning_rate": 1.0614583333333335e-05, "loss": 0.0153, "step": 4610 }, { "epoch": 47.24832214765101, "grad_norm": 0.2933562099933624, "learning_rate": 1.059375e-05, "loss": 0.0151, "step": 4620 }, { "epoch": 47.35059124320869, "grad_norm": 0.3412613272666931, "learning_rate": 1.0572916666666668e-05, "loss": 0.0154, "step": 4630 }, { "epoch": 47.45286033876638, "grad_norm": 0.32954832911491394, "learning_rate": 1.0552083333333333e-05, "loss": 0.0154, "step": 4640 }, { "epoch": 47.555129434324066, "grad_norm": 0.3282068073749542, "learning_rate": 1.0531250000000002e-05, "loss": 0.0152, "step": 4650 }, { "epoch": 47.65739852988175, "grad_norm": 0.33588287234306335, "learning_rate": 1.0510416666666667e-05, "loss": 0.0156, "step": 4660 }, { "epoch": 47.75966762543944, "grad_norm": 0.33600950241088867, "learning_rate": 1.0489583333333335e-05, "loss": 0.0151, "step": 4670 }, { "epoch": 47.86193672099712, "grad_norm": 0.335001140832901, "learning_rate": 1.046875e-05, "loss": 0.0156, "step": 4680 }, { "epoch": 47.96420581655481, "grad_norm": 0.33093252778053284, "learning_rate": 1.0447916666666667e-05, "loss": 0.0158, "step": 4690 }, { "epoch": 48.0664749121125, "grad_norm": 0.32087039947509766, "learning_rate": 1.0427083333333334e-05, "loss": 0.0149, "step": 4700 }, { "epoch": 48.0664749121125, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.274885654449463, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8036, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.137, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.017, "step": 4700 }, { "epoch": 48.168744007670185, "grad_norm": 0.31192752718925476, "learning_rate": 1.040625e-05, "loss": 0.0145, "step": 4710 }, { "epoch": 48.271013103227865, "grad_norm": 0.2866131067276001, "learning_rate": 1.0385416666666667e-05, "loss": 0.0144, "step": 4720 }, { "epoch": 48.37328219878555, "grad_norm": 0.34419262409210205, "learning_rate": 1.0364583333333334e-05, "loss": 0.0145, "step": 4730 }, { "epoch": 48.47555129434324, "grad_norm": 0.33133459091186523, "learning_rate": 1.034375e-05, "loss": 0.0143, "step": 4740 }, { "epoch": 48.57782038990093, "grad_norm": 0.3273964524269104, "learning_rate": 1.0322916666666667e-05, "loss": 0.0146, "step": 4750 }, { "epoch": 48.680089485458616, "grad_norm": 0.3400149643421173, "learning_rate": 1.0302083333333334e-05, "loss": 0.0149, "step": 4760 }, { "epoch": 48.7823585810163, "grad_norm": 0.3198685348033905, "learning_rate": 1.0281250000000001e-05, "loss": 0.0148, "step": 4770 }, { "epoch": 48.884627676573984, "grad_norm": 0.3375503420829773, "learning_rate": 1.0260416666666668e-05, "loss": 0.0153, "step": 4780 }, { "epoch": 48.98689677213167, "grad_norm": 0.3139088749885559, "learning_rate": 1.0239583333333334e-05, "loss": 0.0152, "step": 4790 }, { "epoch": 49.08916586768936, "grad_norm": 0.32369062304496765, "learning_rate": 1.0218750000000001e-05, "loss": 0.0143, "step": 4800 }, { "epoch": 49.08916586768936, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2769861221313477, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7984, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.161, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.02, "step": 4800 }, { "epoch": 49.19143496324704, "grad_norm": 0.3071221113204956, "learning_rate": 1.0197916666666668e-05, "loss": 0.0136, "step": 4810 }, { "epoch": 49.29370405880473, "grad_norm": 0.31288015842437744, "learning_rate": 1.0177083333333335e-05, "loss": 0.0139, "step": 4820 }, { "epoch": 49.395973154362416, "grad_norm": 0.31299805641174316, "learning_rate": 1.0156250000000001e-05, "loss": 0.0137, "step": 4830 }, { "epoch": 49.498242249920104, "grad_norm": 0.3377828001976013, "learning_rate": 1.0135416666666666e-05, "loss": 0.0141, "step": 4840 }, { "epoch": 49.60051134547779, "grad_norm": 0.31973740458488464, "learning_rate": 1.0114583333333335e-05, "loss": 0.0144, "step": 4850 }, { "epoch": 49.70278044103547, "grad_norm": 0.3018786907196045, "learning_rate": 1.009375e-05, "loss": 0.0144, "step": 4860 }, { "epoch": 49.80504953659316, "grad_norm": 0.3308105766773224, "learning_rate": 1.0072916666666668e-05, "loss": 0.0146, "step": 4870 }, { "epoch": 49.90731863215085, "grad_norm": 0.3090561032295227, "learning_rate": 1.0052083333333333e-05, "loss": 0.0145, "step": 4880 }, { "epoch": 50.009587727708535, "grad_norm": 0.33793970942497253, "learning_rate": 1.0031250000000002e-05, "loss": 0.0145, "step": 4890 }, { "epoch": 50.11185682326622, "grad_norm": 0.3269507884979248, "learning_rate": 1.0010416666666667e-05, "loss": 0.0133, "step": 4900 }, { "epoch": 50.11185682326622, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.2938148975372314, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7798, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.247, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.031, "step": 4900 }, { "epoch": 50.2141259188239, "grad_norm": 0.30052992701530457, "learning_rate": 9.989583333333333e-06, "loss": 0.0136, "step": 4910 }, { "epoch": 50.31639501438159, "grad_norm": 0.2977409362792969, "learning_rate": 9.96875e-06, "loss": 0.0134, "step": 4920 }, { "epoch": 50.41866410993928, "grad_norm": 0.3022303283214569, "learning_rate": 9.947916666666667e-06, "loss": 0.0136, "step": 4930 }, { "epoch": 50.520933205496966, "grad_norm": 0.2963425815105438, "learning_rate": 9.927083333333334e-06, "loss": 0.0137, "step": 4940 }, { "epoch": 50.62320230105465, "grad_norm": 0.2862411141395569, "learning_rate": 9.90625e-06, "loss": 0.0134, "step": 4950 }, { "epoch": 50.725471396612335, "grad_norm": 0.32539355754852295, "learning_rate": 9.885416666666667e-06, "loss": 0.0136, "step": 4960 }, { "epoch": 50.82774049217002, "grad_norm": 0.30780264735221863, "learning_rate": 9.864583333333334e-06, "loss": 0.0138, "step": 4970 }, { "epoch": 50.93000958772771, "grad_norm": 0.3183571994304657, "learning_rate": 9.84375e-06, "loss": 0.0136, "step": 4980 }, { "epoch": 51.0322786832854, "grad_norm": 0.28671231865882874, "learning_rate": 9.822916666666667e-06, "loss": 0.0136, "step": 4990 }, { "epoch": 51.13454777884308, "grad_norm": 0.2912745177745819, "learning_rate": 9.802083333333334e-06, "loss": 0.0126, "step": 5000 }, { "epoch": 51.13454777884308, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.312565326690674, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8065, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.124, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.015, "step": 5000 }, { "epoch": 51.236816874400766, "grad_norm": 0.29751163721084595, "learning_rate": 9.78125e-06, "loss": 0.013, "step": 5010 }, { "epoch": 51.339085969958454, "grad_norm": 0.30533960461616516, "learning_rate": 9.760416666666667e-06, "loss": 0.013, "step": 5020 }, { "epoch": 51.44135506551614, "grad_norm": 0.3039548695087433, "learning_rate": 9.739583333333334e-06, "loss": 0.0129, "step": 5030 }, { "epoch": 51.54362416107382, "grad_norm": 0.295386403799057, "learning_rate": 9.71875e-06, "loss": 0.0132, "step": 5040 }, { "epoch": 51.64589325663151, "grad_norm": 0.289328932762146, "learning_rate": 9.697916666666667e-06, "loss": 0.0133, "step": 5050 }, { "epoch": 51.7481623521892, "grad_norm": 0.3070317804813385, "learning_rate": 9.677083333333334e-06, "loss": 0.0131, "step": 5060 }, { "epoch": 51.850431447746885, "grad_norm": 0.28922444581985474, "learning_rate": 9.656250000000001e-06, "loss": 0.0131, "step": 5070 }, { "epoch": 51.95270054330457, "grad_norm": 0.3166629672050476, "learning_rate": 9.635416666666668e-06, "loss": 0.0132, "step": 5080 }, { "epoch": 52.054969638862254, "grad_norm": 0.3101074993610382, "learning_rate": 9.614583333333334e-06, "loss": 0.0126, "step": 5090 }, { "epoch": 52.15723873441994, "grad_norm": 0.318968266248703, "learning_rate": 9.593750000000001e-06, "loss": 0.0127, "step": 5100 }, { "epoch": 52.15723873441994, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3045780658721924, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8544, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.903, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.988, "step": 5100 }, { "epoch": 52.25950782997763, "grad_norm": 0.30769291520118713, "learning_rate": 9.572916666666668e-06, "loss": 0.0122, "step": 5110 }, { "epoch": 52.36177692553532, "grad_norm": 0.28409647941589355, "learning_rate": 9.552083333333335e-06, "loss": 0.0124, "step": 5120 }, { "epoch": 52.464046021093004, "grad_norm": 0.3307039737701416, "learning_rate": 9.531250000000001e-06, "loss": 0.0128, "step": 5130 }, { "epoch": 52.566315116650685, "grad_norm": 0.292473703622818, "learning_rate": 9.510416666666668e-06, "loss": 0.0125, "step": 5140 }, { "epoch": 52.66858421220837, "grad_norm": 0.3042745888233185, "learning_rate": 9.489583333333335e-06, "loss": 0.0126, "step": 5150 }, { "epoch": 52.77085330776606, "grad_norm": 0.29861128330230713, "learning_rate": 9.468750000000001e-06, "loss": 0.0128, "step": 5160 }, { "epoch": 52.87312240332375, "grad_norm": 0.284404993057251, "learning_rate": 9.447916666666668e-06, "loss": 0.0129, "step": 5170 }, { "epoch": 52.97539149888143, "grad_norm": 0.30096396803855896, "learning_rate": 9.427083333333335e-06, "loss": 0.0131, "step": 5180 }, { "epoch": 53.077660594439116, "grad_norm": 0.3048815131187439, "learning_rate": 9.406250000000002e-06, "loss": 0.012, "step": 5190 }, { "epoch": 53.179929689996804, "grad_norm": 0.2931344509124756, "learning_rate": 9.385416666666668e-06, "loss": 0.0118, "step": 5200 }, { "epoch": 53.179929689996804, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.321913480758667, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.8669, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 95.845, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 11.981, "step": 5200 }, { "epoch": 53.28219878555449, "grad_norm": 0.29786059260368347, "learning_rate": 9.364583333333333e-06, "loss": 0.0116, "step": 5210 }, { "epoch": 53.38446788111218, "grad_norm": 0.2869616150856018, "learning_rate": 9.34375e-06, "loss": 0.0119, "step": 5220 }, { "epoch": 53.48673697666986, "grad_norm": 0.3017300069332123, "learning_rate": 9.322916666666667e-06, "loss": 0.0121, "step": 5230 }, { "epoch": 53.58900607222755, "grad_norm": 0.3326238691806793, "learning_rate": 9.302083333333334e-06, "loss": 0.0123, "step": 5240 }, { "epoch": 53.691275167785236, "grad_norm": 0.30569109320640564, "learning_rate": 9.28125e-06, "loss": 0.0125, "step": 5250 }, { "epoch": 53.79354426334292, "grad_norm": 0.3061717748641968, "learning_rate": 9.260416666666667e-06, "loss": 0.0123, "step": 5260 }, { "epoch": 53.895813358900604, "grad_norm": 0.281955748796463, "learning_rate": 9.239583333333334e-06, "loss": 0.0121, "step": 5270 }, { "epoch": 53.99808245445829, "grad_norm": 0.2920975983142853, "learning_rate": 9.21875e-06, "loss": 0.0124, "step": 5280 }, { "epoch": 54.10035155001598, "grad_norm": 0.31395605206489563, "learning_rate": 9.197916666666667e-06, "loss": 0.0116, "step": 5290 }, { "epoch": 54.20262064557367, "grad_norm": 0.3078666627407074, "learning_rate": 9.177083333333334e-06, "loss": 0.0117, "step": 5300 }, { "epoch": 54.20262064557367, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3260598182678223, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7898, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.201, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.025, "step": 5300 }, { "epoch": 54.304889741131355, "grad_norm": 0.2747853994369507, "learning_rate": 9.15625e-06, "loss": 0.0113, "step": 5310 }, { "epoch": 54.407158836689035, "grad_norm": 0.3068675398826599, "learning_rate": 9.135416666666667e-06, "loss": 0.0115, "step": 5320 }, { "epoch": 54.50942793224672, "grad_norm": 0.29503345489501953, "learning_rate": 9.114583333333334e-06, "loss": 0.0116, "step": 5330 }, { "epoch": 54.61169702780441, "grad_norm": 0.27636656165122986, "learning_rate": 9.09375e-06, "loss": 0.0114, "step": 5340 }, { "epoch": 54.7139661233621, "grad_norm": 0.313203901052475, "learning_rate": 9.072916666666668e-06, "loss": 0.012, "step": 5350 }, { "epoch": 54.816235218919786, "grad_norm": 0.30984997749328613, "learning_rate": 9.052083333333334e-06, "loss": 0.012, "step": 5360 }, { "epoch": 54.91850431447747, "grad_norm": 0.27004456520080566, "learning_rate": 9.031250000000001e-06, "loss": 0.0118, "step": 5370 }, { "epoch": 55.020773410035154, "grad_norm": 0.278777152299881, "learning_rate": 9.010416666666668e-06, "loss": 0.0117, "step": 5380 }, { "epoch": 55.12304250559284, "grad_norm": 0.3070172965526581, "learning_rate": 8.989583333333334e-06, "loss": 0.011, "step": 5390 }, { "epoch": 55.22531160115053, "grad_norm": 0.2756708562374115, "learning_rate": 8.968750000000001e-06, "loss": 0.011, "step": 5400 }, { "epoch": 55.22531160115053, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.33424711227417, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7849, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.224, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.028, "step": 5400 }, { "epoch": 55.32758069670821, "grad_norm": 0.28067439794540405, "learning_rate": 8.947916666666668e-06, "loss": 0.011, "step": 5410 }, { "epoch": 55.4298497922659, "grad_norm": 0.2702157199382782, "learning_rate": 8.927083333333335e-06, "loss": 0.0112, "step": 5420 }, { "epoch": 55.532118887823586, "grad_norm": 0.30056026577949524, "learning_rate": 8.906250000000001e-06, "loss": 0.0115, "step": 5430 }, { "epoch": 55.634387983381274, "grad_norm": 0.29118791222572327, "learning_rate": 8.885416666666668e-06, "loss": 0.0113, "step": 5440 }, { "epoch": 55.73665707893896, "grad_norm": 0.3045700490474701, "learning_rate": 8.864583333333335e-06, "loss": 0.0115, "step": 5450 }, { "epoch": 55.83892617449664, "grad_norm": 0.28140169382095337, "learning_rate": 8.843750000000002e-06, "loss": 0.0113, "step": 5460 }, { "epoch": 55.94119527005433, "grad_norm": 0.2764737010002136, "learning_rate": 8.822916666666668e-06, "loss": 0.0118, "step": 5470 }, { "epoch": 56.04346436561202, "grad_norm": 0.28735798597335815, "learning_rate": 8.802083333333335e-06, "loss": 0.0114, "step": 5480 }, { "epoch": 56.145733461169705, "grad_norm": 0.28290146589279175, "learning_rate": 8.781250000000002e-06, "loss": 0.0107, "step": 5490 }, { "epoch": 56.248002556727386, "grad_norm": 0.2917637526988983, "learning_rate": 8.760416666666668e-06, "loss": 0.0108, "step": 5500 }, { "epoch": 56.248002556727386, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3639163970947266, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7163, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.542, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.068, "step": 5500 }, { "epoch": 56.35027165228507, "grad_norm": 0.31157800555229187, "learning_rate": 8.739583333333333e-06, "loss": 0.0109, "step": 5510 }, { "epoch": 56.45254074784276, "grad_norm": 0.26494109630584717, "learning_rate": 8.71875e-06, "loss": 0.0108, "step": 5520 }, { "epoch": 56.55480984340045, "grad_norm": 0.2839730381965637, "learning_rate": 8.697916666666667e-06, "loss": 0.011, "step": 5530 }, { "epoch": 56.657078938958136, "grad_norm": 0.27168846130371094, "learning_rate": 8.677083333333334e-06, "loss": 0.0111, "step": 5540 }, { "epoch": 56.75934803451582, "grad_norm": 0.2758902311325073, "learning_rate": 8.65625e-06, "loss": 0.0109, "step": 5550 }, { "epoch": 56.861617130073505, "grad_norm": 0.29986515641212463, "learning_rate": 8.635416666666667e-06, "loss": 0.0113, "step": 5560 }, { "epoch": 56.96388622563119, "grad_norm": 0.2895634174346924, "learning_rate": 8.614583333333334e-06, "loss": 0.0112, "step": 5570 }, { "epoch": 57.06615532118888, "grad_norm": 0.26848530769348145, "learning_rate": 8.59375e-06, "loss": 0.0105, "step": 5580 }, { "epoch": 57.16842441674657, "grad_norm": 0.2801918685436249, "learning_rate": 8.572916666666667e-06, "loss": 0.0106, "step": 5590 }, { "epoch": 57.27069351230425, "grad_norm": 0.2772628366947174, "learning_rate": 8.552083333333334e-06, "loss": 0.0105, "step": 5600 }, { "epoch": 57.27069351230425, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3490209579467773, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7134, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.556, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.069, "step": 5600 }, { "epoch": 57.372962607861936, "grad_norm": 0.26178139448165894, "learning_rate": 8.53125e-06, "loss": 0.0103, "step": 5610 }, { "epoch": 57.475231703419624, "grad_norm": 0.27127423882484436, "learning_rate": 8.510416666666667e-06, "loss": 0.0105, "step": 5620 }, { "epoch": 57.57750079897731, "grad_norm": 0.2728956639766693, "learning_rate": 8.489583333333334e-06, "loss": 0.0104, "step": 5630 }, { "epoch": 57.67976989453499, "grad_norm": 0.26934826374053955, "learning_rate": 8.468750000000001e-06, "loss": 0.0107, "step": 5640 }, { "epoch": 57.78203899009268, "grad_norm": 0.30065199732780457, "learning_rate": 8.447916666666668e-06, "loss": 0.0108, "step": 5650 }, { "epoch": 57.88430808565037, "grad_norm": 0.29894405603408813, "learning_rate": 8.427083333333334e-06, "loss": 0.0104, "step": 5660 }, { "epoch": 57.986577181208055, "grad_norm": 0.26421451568603516, "learning_rate": 8.406250000000001e-06, "loss": 0.0105, "step": 5670 }, { "epoch": 58.08884627676574, "grad_norm": 0.27239277958869934, "learning_rate": 8.385416666666668e-06, "loss": 0.0099, "step": 5680 }, { "epoch": 58.191115372323424, "grad_norm": 0.27293631434440613, "learning_rate": 8.364583333333334e-06, "loss": 0.01, "step": 5690 }, { "epoch": 58.29338446788111, "grad_norm": 0.251788467168808, "learning_rate": 8.343750000000001e-06, "loss": 0.0103, "step": 5700 }, { "epoch": 58.29338446788111, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.360133647918701, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7658, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.312, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.039, "step": 5700 }, { "epoch": 58.3956535634388, "grad_norm": 0.2742460072040558, "learning_rate": 8.322916666666668e-06, "loss": 0.01, "step": 5710 }, { "epoch": 58.49792265899649, "grad_norm": 0.2849868834018707, "learning_rate": 8.302083333333335e-06, "loss": 0.0101, "step": 5720 }, { "epoch": 58.60019175455417, "grad_norm": 0.2823048233985901, "learning_rate": 8.281250000000001e-06, "loss": 0.0101, "step": 5730 }, { "epoch": 58.702460850111855, "grad_norm": 0.2816413342952728, "learning_rate": 8.260416666666668e-06, "loss": 0.0102, "step": 5740 }, { "epoch": 58.80472994566954, "grad_norm": 0.3086257576942444, "learning_rate": 8.239583333333335e-06, "loss": 0.0102, "step": 5750 }, { "epoch": 58.90699904122723, "grad_norm": 0.31176137924194336, "learning_rate": 8.218750000000002e-06, "loss": 0.0102, "step": 5760 }, { "epoch": 59.00926813678492, "grad_norm": 0.30020809173583984, "learning_rate": 8.197916666666668e-06, "loss": 0.0103, "step": 5770 }, { "epoch": 59.1115372323426, "grad_norm": 0.27167460322380066, "learning_rate": 8.177083333333335e-06, "loss": 0.0096, "step": 5780 }, { "epoch": 59.213806327900286, "grad_norm": 0.29027628898620605, "learning_rate": 8.156250000000002e-06, "loss": 0.0099, "step": 5790 }, { "epoch": 59.316075423457974, "grad_norm": 0.2502238154411316, "learning_rate": 8.135416666666668e-06, "loss": 0.0097, "step": 5800 }, { "epoch": 59.316075423457974, "eval_silk-road/alpaca-data-gpt4-chinese_loss": 2.3818840980529785, "eval_silk-road/alpaca-data-gpt4-chinese_runtime": 20.7922, "eval_silk-road/alpaca-data-gpt4-chinese_samples_per_second": 96.19, "eval_silk-road/alpaca-data-gpt4-chinese_steps_per_second": 12.024, "step": 5800 } ], "logging_steps": 10, "max_steps": 9700, "num_input_tokens_seen": 0, "num_train_epochs": 100, "save_steps": 200, "total_flos": 7.772644061346693e+18, "train_batch_size": 16, "trial_name": null, "trial_params": null }