GaetanMichelet's picture
Model save
4dbaa51 verified
raw
history blame contribute delete
No virus
46.1 kB
{
"best_metric": 0.48128727078437805,
"best_model_checkpoint": "data/Llama-31-8B_task-3_180-samples_config-3/checkpoint-340",
"epoch": 27.0,
"eval_steps": 500,
"global_step": 459,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.058823529411764705,
"grad_norm": 2.980715036392212,
"learning_rate": 3.9215686274509804e-08,
"loss": 2.6305,
"step": 1
},
{
"epoch": 0.11764705882352941,
"grad_norm": 1.5815191268920898,
"learning_rate": 7.843137254901961e-08,
"loss": 2.0227,
"step": 2
},
{
"epoch": 0.23529411764705882,
"grad_norm": 3.0497844219207764,
"learning_rate": 1.5686274509803921e-07,
"loss": 2.8101,
"step": 4
},
{
"epoch": 0.35294117647058826,
"grad_norm": 1.5842593908309937,
"learning_rate": 2.3529411764705883e-07,
"loss": 2.5918,
"step": 6
},
{
"epoch": 0.47058823529411764,
"grad_norm": 2.9371345043182373,
"learning_rate": 3.1372549019607843e-07,
"loss": 2.3864,
"step": 8
},
{
"epoch": 0.5882352941176471,
"grad_norm": 2.9547200202941895,
"learning_rate": 3.921568627450981e-07,
"loss": 2.6168,
"step": 10
},
{
"epoch": 0.7058823529411765,
"grad_norm": 2.5302109718322754,
"learning_rate": 4.7058823529411767e-07,
"loss": 2.2036,
"step": 12
},
{
"epoch": 0.8235294117647058,
"grad_norm": 3.466365337371826,
"learning_rate": 5.490196078431373e-07,
"loss": 2.7139,
"step": 14
},
{
"epoch": 0.9411764705882353,
"grad_norm": 1.7582124471664429,
"learning_rate": 6.274509803921569e-07,
"loss": 2.3708,
"step": 16
},
{
"epoch": 1.0,
"eval_loss": 2.4981932640075684,
"eval_runtime": 31.8694,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 17
},
{
"epoch": 1.0588235294117647,
"grad_norm": 1.3897138833999634,
"learning_rate": 7.058823529411766e-07,
"loss": 2.2296,
"step": 18
},
{
"epoch": 1.1764705882352942,
"grad_norm": 2.683563470840454,
"learning_rate": 7.843137254901962e-07,
"loss": 2.5726,
"step": 20
},
{
"epoch": 1.2941176470588236,
"grad_norm": 2.160468816757202,
"learning_rate": 8.627450980392157e-07,
"loss": 2.4421,
"step": 22
},
{
"epoch": 1.4117647058823528,
"grad_norm": 3.092349052429199,
"learning_rate": 9.411764705882353e-07,
"loss": 2.7441,
"step": 24
},
{
"epoch": 1.5294117647058822,
"grad_norm": 3.0546014308929443,
"learning_rate": 1.019607843137255e-06,
"loss": 2.6233,
"step": 26
},
{
"epoch": 1.6470588235294117,
"grad_norm": 2.6419599056243896,
"learning_rate": 1.0980392156862745e-06,
"loss": 2.6529,
"step": 28
},
{
"epoch": 1.7647058823529411,
"grad_norm": 1.2905603647232056,
"learning_rate": 1.1764705882352942e-06,
"loss": 2.1026,
"step": 30
},
{
"epoch": 1.8823529411764706,
"grad_norm": 2.6626884937286377,
"learning_rate": 1.2549019607843137e-06,
"loss": 2.5973,
"step": 32
},
{
"epoch": 2.0,
"grad_norm": 3.445452928543091,
"learning_rate": 1.3333333333333334e-06,
"loss": 2.4065,
"step": 34
},
{
"epoch": 2.0,
"eval_loss": 2.439713954925537,
"eval_runtime": 31.87,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 34
},
{
"epoch": 2.1176470588235294,
"grad_norm": 2.4814882278442383,
"learning_rate": 1.4117647058823531e-06,
"loss": 2.3601,
"step": 36
},
{
"epoch": 2.235294117647059,
"grad_norm": 3.0113446712493896,
"learning_rate": 1.4901960784313726e-06,
"loss": 2.3377,
"step": 38
},
{
"epoch": 2.3529411764705883,
"grad_norm": 2.5856285095214844,
"learning_rate": 1.5686274509803923e-06,
"loss": 2.24,
"step": 40
},
{
"epoch": 2.4705882352941178,
"grad_norm": 3.1034915447235107,
"learning_rate": 1.6470588235294118e-06,
"loss": 2.5231,
"step": 42
},
{
"epoch": 2.588235294117647,
"grad_norm": 3.499271869659424,
"learning_rate": 1.7254901960784315e-06,
"loss": 2.4463,
"step": 44
},
{
"epoch": 2.7058823529411766,
"grad_norm": 2.990689992904663,
"learning_rate": 1.8039215686274512e-06,
"loss": 2.4596,
"step": 46
},
{
"epoch": 2.8235294117647056,
"grad_norm": 2.712009906768799,
"learning_rate": 1.8823529411764707e-06,
"loss": 2.2987,
"step": 48
},
{
"epoch": 2.9411764705882355,
"grad_norm": 3.5523183345794678,
"learning_rate": 1.96078431372549e-06,
"loss": 2.3549,
"step": 50
},
{
"epoch": 3.0,
"eval_loss": 2.31467604637146,
"eval_runtime": 31.8743,
"eval_samples_per_second": 1.129,
"eval_steps_per_second": 1.129,
"step": 51
},
{
"epoch": 3.0588235294117645,
"grad_norm": 3.0498082637786865,
"learning_rate": 2.03921568627451e-06,
"loss": 2.4567,
"step": 52
},
{
"epoch": 3.176470588235294,
"grad_norm": 2.239677906036377,
"learning_rate": 2.1176470588235296e-06,
"loss": 2.0235,
"step": 54
},
{
"epoch": 3.2941176470588234,
"grad_norm": 2.1120095252990723,
"learning_rate": 2.196078431372549e-06,
"loss": 2.2331,
"step": 56
},
{
"epoch": 3.411764705882353,
"grad_norm": 3.549705743789673,
"learning_rate": 2.274509803921569e-06,
"loss": 2.1111,
"step": 58
},
{
"epoch": 3.5294117647058822,
"grad_norm": 4.140646934509277,
"learning_rate": 2.3529411764705885e-06,
"loss": 2.4917,
"step": 60
},
{
"epoch": 3.6470588235294117,
"grad_norm": 3.1374101638793945,
"learning_rate": 2.431372549019608e-06,
"loss": 2.2287,
"step": 62
},
{
"epoch": 3.764705882352941,
"grad_norm": 2.198146104812622,
"learning_rate": 2.5098039215686274e-06,
"loss": 1.9659,
"step": 64
},
{
"epoch": 3.8823529411764706,
"grad_norm": 2.7118916511535645,
"learning_rate": 2.5882352941176473e-06,
"loss": 2.3314,
"step": 66
},
{
"epoch": 4.0,
"grad_norm": 3.456117868423462,
"learning_rate": 2.666666666666667e-06,
"loss": 2.0578,
"step": 68
},
{
"epoch": 4.0,
"eval_loss": 2.085047483444214,
"eval_runtime": 31.8792,
"eval_samples_per_second": 1.129,
"eval_steps_per_second": 1.129,
"step": 68
},
{
"epoch": 4.117647058823529,
"grad_norm": 2.2668004035949707,
"learning_rate": 2.7450980392156867e-06,
"loss": 2.1099,
"step": 70
},
{
"epoch": 4.235294117647059,
"grad_norm": 2.0983450412750244,
"learning_rate": 2.8235294117647062e-06,
"loss": 2.0241,
"step": 72
},
{
"epoch": 4.352941176470588,
"grad_norm": 1.6723711490631104,
"learning_rate": 2.901960784313726e-06,
"loss": 1.8905,
"step": 74
},
{
"epoch": 4.470588235294118,
"grad_norm": 3.0844905376434326,
"learning_rate": 2.980392156862745e-06,
"loss": 1.8637,
"step": 76
},
{
"epoch": 4.588235294117647,
"grad_norm": 3.0651304721832275,
"learning_rate": 3.058823529411765e-06,
"loss": 1.8883,
"step": 78
},
{
"epoch": 4.705882352941177,
"grad_norm": 2.8657171726226807,
"learning_rate": 3.1372549019607846e-06,
"loss": 1.8309,
"step": 80
},
{
"epoch": 4.823529411764706,
"grad_norm": 2.580749750137329,
"learning_rate": 3.2156862745098045e-06,
"loss": 1.7815,
"step": 82
},
{
"epoch": 4.9411764705882355,
"grad_norm": 2.4597244262695312,
"learning_rate": 3.2941176470588236e-06,
"loss": 1.8089,
"step": 84
},
{
"epoch": 5.0,
"eval_loss": 1.7080037593841553,
"eval_runtime": 31.8717,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 85
},
{
"epoch": 5.0588235294117645,
"grad_norm": 1.954211950302124,
"learning_rate": 3.3725490196078435e-06,
"loss": 1.6625,
"step": 86
},
{
"epoch": 5.176470588235294,
"grad_norm": 1.9714858531951904,
"learning_rate": 3.450980392156863e-06,
"loss": 1.5611,
"step": 88
},
{
"epoch": 5.294117647058823,
"grad_norm": 1.9127044677734375,
"learning_rate": 3.529411764705883e-06,
"loss": 1.5702,
"step": 90
},
{
"epoch": 5.411764705882353,
"grad_norm": 4.0653977394104,
"learning_rate": 3.6078431372549024e-06,
"loss": 1.5627,
"step": 92
},
{
"epoch": 5.529411764705882,
"grad_norm": 1.8669339418411255,
"learning_rate": 3.6862745098039223e-06,
"loss": 1.4645,
"step": 94
},
{
"epoch": 5.647058823529412,
"grad_norm": 2.6541359424591064,
"learning_rate": 3.7647058823529414e-06,
"loss": 1.3323,
"step": 96
},
{
"epoch": 5.764705882352941,
"grad_norm": 1.7167555093765259,
"learning_rate": 3.843137254901962e-06,
"loss": 1.3833,
"step": 98
},
{
"epoch": 5.882352941176471,
"grad_norm": 2.953925132751465,
"learning_rate": 3.92156862745098e-06,
"loss": 1.333,
"step": 100
},
{
"epoch": 6.0,
"grad_norm": 1.2611362934112549,
"learning_rate": 4.000000000000001e-06,
"loss": 1.3018,
"step": 102
},
{
"epoch": 6.0,
"eval_loss": 1.2346911430358887,
"eval_runtime": 31.8703,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 102
},
{
"epoch": 6.117647058823529,
"grad_norm": 1.9918830394744873,
"learning_rate": 4.07843137254902e-06,
"loss": 1.1196,
"step": 104
},
{
"epoch": 6.235294117647059,
"grad_norm": 1.8971362113952637,
"learning_rate": 4.15686274509804e-06,
"loss": 1.1518,
"step": 106
},
{
"epoch": 6.352941176470588,
"grad_norm": 2.2451581954956055,
"learning_rate": 4.235294117647059e-06,
"loss": 1.0684,
"step": 108
},
{
"epoch": 6.470588235294118,
"grad_norm": 2.1816394329071045,
"learning_rate": 4.313725490196079e-06,
"loss": 0.9634,
"step": 110
},
{
"epoch": 6.588235294117647,
"grad_norm": 1.7623378038406372,
"learning_rate": 4.392156862745098e-06,
"loss": 0.9115,
"step": 112
},
{
"epoch": 6.705882352941177,
"grad_norm": 2.0872268676757812,
"learning_rate": 4.4705882352941184e-06,
"loss": 0.7509,
"step": 114
},
{
"epoch": 6.823529411764706,
"grad_norm": 1.5808390378952026,
"learning_rate": 4.549019607843138e-06,
"loss": 0.8618,
"step": 116
},
{
"epoch": 6.9411764705882355,
"grad_norm": 1.1898610591888428,
"learning_rate": 4.627450980392157e-06,
"loss": 1.0212,
"step": 118
},
{
"epoch": 7.0,
"eval_loss": 0.8016352653503418,
"eval_runtime": 31.8718,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 119
},
{
"epoch": 7.0588235294117645,
"grad_norm": 1.153451919555664,
"learning_rate": 4.705882352941177e-06,
"loss": 0.9469,
"step": 120
},
{
"epoch": 7.176470588235294,
"grad_norm": 1.1448363065719604,
"learning_rate": 4.784313725490196e-06,
"loss": 0.9065,
"step": 122
},
{
"epoch": 7.294117647058823,
"grad_norm": 1.0351287126541138,
"learning_rate": 4.862745098039216e-06,
"loss": 0.6538,
"step": 124
},
{
"epoch": 7.411764705882353,
"grad_norm": 0.9266816973686218,
"learning_rate": 4.941176470588236e-06,
"loss": 0.7569,
"step": 126
},
{
"epoch": 7.529411764705882,
"grad_norm": 0.869621992111206,
"learning_rate": 5.019607843137255e-06,
"loss": 0.6096,
"step": 128
},
{
"epoch": 7.647058823529412,
"grad_norm": 0.8324370980262756,
"learning_rate": 5.098039215686274e-06,
"loss": 0.4993,
"step": 130
},
{
"epoch": 7.764705882352941,
"grad_norm": 0.8301700353622437,
"learning_rate": 5.176470588235295e-06,
"loss": 0.4293,
"step": 132
},
{
"epoch": 7.882352941176471,
"grad_norm": 0.3732304275035858,
"learning_rate": 5.254901960784314e-06,
"loss": 0.632,
"step": 134
},
{
"epoch": 8.0,
"grad_norm": 0.5899373888969421,
"learning_rate": 5.333333333333334e-06,
"loss": 0.4899,
"step": 136
},
{
"epoch": 8.0,
"eval_loss": 0.6475194096565247,
"eval_runtime": 31.8748,
"eval_samples_per_second": 1.129,
"eval_steps_per_second": 1.129,
"step": 136
},
{
"epoch": 8.117647058823529,
"grad_norm": 0.6668117642402649,
"learning_rate": 5.411764705882353e-06,
"loss": 0.468,
"step": 138
},
{
"epoch": 8.235294117647058,
"grad_norm": 1.0659523010253906,
"learning_rate": 5.4901960784313735e-06,
"loss": 0.6599,
"step": 140
},
{
"epoch": 8.352941176470589,
"grad_norm": 0.787837028503418,
"learning_rate": 5.568627450980393e-06,
"loss": 0.8816,
"step": 142
},
{
"epoch": 8.470588235294118,
"grad_norm": 0.9166098237037659,
"learning_rate": 5.6470588235294125e-06,
"loss": 0.4608,
"step": 144
},
{
"epoch": 8.588235294117647,
"grad_norm": 0.8993943929672241,
"learning_rate": 5.725490196078431e-06,
"loss": 0.6013,
"step": 146
},
{
"epoch": 8.705882352941176,
"grad_norm": 0.5819871425628662,
"learning_rate": 5.803921568627452e-06,
"loss": 0.2645,
"step": 148
},
{
"epoch": 8.823529411764707,
"grad_norm": 0.6202341318130493,
"learning_rate": 5.882352941176471e-06,
"loss": 0.5653,
"step": 150
},
{
"epoch": 8.941176470588236,
"grad_norm": 0.737612247467041,
"learning_rate": 5.96078431372549e-06,
"loss": 0.6106,
"step": 152
},
{
"epoch": 9.0,
"eval_loss": 0.5890491008758545,
"eval_runtime": 31.8673,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 153
},
{
"epoch": 9.058823529411764,
"grad_norm": 0.8201161623001099,
"learning_rate": 6.03921568627451e-06,
"loss": 0.556,
"step": 154
},
{
"epoch": 9.176470588235293,
"grad_norm": 0.5118715167045593,
"learning_rate": 6.11764705882353e-06,
"loss": 0.521,
"step": 156
},
{
"epoch": 9.294117647058824,
"grad_norm": 2.472463369369507,
"learning_rate": 6.19607843137255e-06,
"loss": 0.5874,
"step": 158
},
{
"epoch": 9.411764705882353,
"grad_norm": 0.6849164366722107,
"learning_rate": 6.274509803921569e-06,
"loss": 0.5052,
"step": 160
},
{
"epoch": 9.529411764705882,
"grad_norm": 0.5705221891403198,
"learning_rate": 6.352941176470589e-06,
"loss": 0.4448,
"step": 162
},
{
"epoch": 9.647058823529411,
"grad_norm": 0.9986915588378906,
"learning_rate": 6.431372549019609e-06,
"loss": 0.4622,
"step": 164
},
{
"epoch": 9.764705882352942,
"grad_norm": 0.5936263203620911,
"learning_rate": 6.5098039215686285e-06,
"loss": 0.5221,
"step": 166
},
{
"epoch": 9.882352941176471,
"grad_norm": 0.4011266231536865,
"learning_rate": 6.588235294117647e-06,
"loss": 0.3669,
"step": 168
},
{
"epoch": 10.0,
"grad_norm": 1.4033674001693726,
"learning_rate": 6.666666666666667e-06,
"loss": 0.5388,
"step": 170
},
{
"epoch": 10.0,
"eval_loss": 0.5729286670684814,
"eval_runtime": 31.8736,
"eval_samples_per_second": 1.129,
"eval_steps_per_second": 1.129,
"step": 170
},
{
"epoch": 10.117647058823529,
"grad_norm": 0.7268795371055603,
"learning_rate": 6.745098039215687e-06,
"loss": 0.4123,
"step": 172
},
{
"epoch": 10.235294117647058,
"grad_norm": 0.534287691116333,
"learning_rate": 6.8235294117647065e-06,
"loss": 0.3577,
"step": 174
},
{
"epoch": 10.352941176470589,
"grad_norm": 0.46323052048683167,
"learning_rate": 6.901960784313726e-06,
"loss": 0.3157,
"step": 176
},
{
"epoch": 10.470588235294118,
"grad_norm": 0.4801871180534363,
"learning_rate": 6.9803921568627454e-06,
"loss": 0.5767,
"step": 178
},
{
"epoch": 10.588235294117647,
"grad_norm": 0.47244688868522644,
"learning_rate": 7.058823529411766e-06,
"loss": 0.5139,
"step": 180
},
{
"epoch": 10.705882352941176,
"grad_norm": 0.41083067655563354,
"learning_rate": 7.137254901960785e-06,
"loss": 0.5551,
"step": 182
},
{
"epoch": 10.823529411764707,
"grad_norm": 0.6211682558059692,
"learning_rate": 7.215686274509805e-06,
"loss": 0.5971,
"step": 184
},
{
"epoch": 10.941176470588236,
"grad_norm": 0.4872354567050934,
"learning_rate": 7.294117647058823e-06,
"loss": 0.7245,
"step": 186
},
{
"epoch": 11.0,
"eval_loss": 0.5584969520568848,
"eval_runtime": 31.8629,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 187
},
{
"epoch": 11.058823529411764,
"grad_norm": 0.41652852296829224,
"learning_rate": 7.372549019607845e-06,
"loss": 0.3599,
"step": 188
},
{
"epoch": 11.176470588235293,
"grad_norm": 0.5096337795257568,
"learning_rate": 7.450980392156863e-06,
"loss": 0.5491,
"step": 190
},
{
"epoch": 11.294117647058824,
"grad_norm": 0.5131371021270752,
"learning_rate": 7.529411764705883e-06,
"loss": 0.6327,
"step": 192
},
{
"epoch": 11.411764705882353,
"grad_norm": 0.33539438247680664,
"learning_rate": 7.607843137254902e-06,
"loss": 0.4431,
"step": 194
},
{
"epoch": 11.529411764705882,
"grad_norm": 1.0883010625839233,
"learning_rate": 7.686274509803923e-06,
"loss": 0.5964,
"step": 196
},
{
"epoch": 11.647058823529411,
"grad_norm": 0.13403330743312836,
"learning_rate": 7.764705882352941e-06,
"loss": 0.1832,
"step": 198
},
{
"epoch": 11.764705882352942,
"grad_norm": 0.38453081250190735,
"learning_rate": 7.84313725490196e-06,
"loss": 0.5463,
"step": 200
},
{
"epoch": 11.882352941176471,
"grad_norm": 0.4325025677680969,
"learning_rate": 7.92156862745098e-06,
"loss": 0.3552,
"step": 202
},
{
"epoch": 12.0,
"grad_norm": 0.26013830304145813,
"learning_rate": 8.000000000000001e-06,
"loss": 0.3568,
"step": 204
},
{
"epoch": 12.0,
"eval_loss": 0.5533129572868347,
"eval_runtime": 31.8552,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 204
},
{
"epoch": 12.117647058823529,
"grad_norm": 0.5007168054580688,
"learning_rate": 8.07843137254902e-06,
"loss": 0.3615,
"step": 206
},
{
"epoch": 12.235294117647058,
"grad_norm": 0.4056757986545563,
"learning_rate": 8.15686274509804e-06,
"loss": 0.7152,
"step": 208
},
{
"epoch": 12.352941176470589,
"grad_norm": 0.2833678424358368,
"learning_rate": 8.23529411764706e-06,
"loss": 0.4289,
"step": 210
},
{
"epoch": 12.470588235294118,
"grad_norm": 0.3110792934894562,
"learning_rate": 8.31372549019608e-06,
"loss": 0.4519,
"step": 212
},
{
"epoch": 12.588235294117647,
"grad_norm": 0.3763163387775421,
"learning_rate": 8.392156862745099e-06,
"loss": 0.405,
"step": 214
},
{
"epoch": 12.705882352941176,
"grad_norm": 0.24176886677742004,
"learning_rate": 8.470588235294118e-06,
"loss": 0.4032,
"step": 216
},
{
"epoch": 12.823529411764707,
"grad_norm": 0.3594723343849182,
"learning_rate": 8.549019607843138e-06,
"loss": 0.3982,
"step": 218
},
{
"epoch": 12.941176470588236,
"grad_norm": 0.50929194688797,
"learning_rate": 8.627450980392157e-06,
"loss": 0.4165,
"step": 220
},
{
"epoch": 13.0,
"eval_loss": 0.5352616310119629,
"eval_runtime": 31.8489,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 221
},
{
"epoch": 13.058823529411764,
"grad_norm": 0.32427483797073364,
"learning_rate": 8.705882352941177e-06,
"loss": 0.49,
"step": 222
},
{
"epoch": 13.176470588235293,
"grad_norm": 0.24497142434120178,
"learning_rate": 8.784313725490196e-06,
"loss": 0.3279,
"step": 224
},
{
"epoch": 13.294117647058824,
"grad_norm": 0.34544771909713745,
"learning_rate": 8.862745098039216e-06,
"loss": 0.4668,
"step": 226
},
{
"epoch": 13.411764705882353,
"grad_norm": 0.2216784805059433,
"learning_rate": 8.941176470588237e-06,
"loss": 0.3077,
"step": 228
},
{
"epoch": 13.529411764705882,
"grad_norm": 0.2850175201892853,
"learning_rate": 9.019607843137256e-06,
"loss": 0.3767,
"step": 230
},
{
"epoch": 13.647058823529411,
"grad_norm": 0.4109225571155548,
"learning_rate": 9.098039215686276e-06,
"loss": 0.4402,
"step": 232
},
{
"epoch": 13.764705882352942,
"grad_norm": 0.24095743894577026,
"learning_rate": 9.176470588235294e-06,
"loss": 0.3823,
"step": 234
},
{
"epoch": 13.882352941176471,
"grad_norm": 0.5017974972724915,
"learning_rate": 9.254901960784315e-06,
"loss": 0.5895,
"step": 236
},
{
"epoch": 14.0,
"grad_norm": 0.45756474137306213,
"learning_rate": 9.333333333333334e-06,
"loss": 0.6226,
"step": 238
},
{
"epoch": 14.0,
"eval_loss": 0.5420221090316772,
"eval_runtime": 31.849,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 238
},
{
"epoch": 14.117647058823529,
"grad_norm": 0.4417901039123535,
"learning_rate": 9.411764705882354e-06,
"loss": 0.545,
"step": 240
},
{
"epoch": 14.235294117647058,
"grad_norm": 0.3756580054759979,
"learning_rate": 9.490196078431373e-06,
"loss": 0.4739,
"step": 242
},
{
"epoch": 14.352941176470589,
"grad_norm": 0.4785820543766022,
"learning_rate": 9.568627450980393e-06,
"loss": 0.4149,
"step": 244
},
{
"epoch": 14.470588235294118,
"grad_norm": 0.444455087184906,
"learning_rate": 9.647058823529412e-06,
"loss": 0.508,
"step": 246
},
{
"epoch": 14.588235294117647,
"grad_norm": 0.3459950387477875,
"learning_rate": 9.725490196078432e-06,
"loss": 0.3752,
"step": 248
},
{
"epoch": 14.705882352941176,
"grad_norm": 0.276597797870636,
"learning_rate": 9.803921568627451e-06,
"loss": 0.3527,
"step": 250
},
{
"epoch": 14.823529411764707,
"grad_norm": 0.31315046548843384,
"learning_rate": 9.882352941176472e-06,
"loss": 0.396,
"step": 252
},
{
"epoch": 14.941176470588236,
"grad_norm": 0.009531126357614994,
"learning_rate": 9.960784313725492e-06,
"loss": 0.3284,
"step": 254
},
{
"epoch": 15.0,
"eval_loss": 0.5025795102119446,
"eval_runtime": 31.8517,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 255
},
{
"epoch": 15.058823529411764,
"grad_norm": 0.5613787174224854,
"learning_rate": 9.999995315380667e-06,
"loss": 0.502,
"step": 256
},
{
"epoch": 15.176470588235293,
"grad_norm": 0.467040091753006,
"learning_rate": 9.99995783847866e-06,
"loss": 0.4182,
"step": 258
},
{
"epoch": 15.294117647058824,
"grad_norm": 0.41092485189437866,
"learning_rate": 9.999882884955554e-06,
"loss": 0.245,
"step": 260
},
{
"epoch": 15.411764705882353,
"grad_norm": 0.36648833751678467,
"learning_rate": 9.99977045537315e-06,
"loss": 0.3851,
"step": 262
},
{
"epoch": 15.529411764705882,
"grad_norm": 0.31992214918136597,
"learning_rate": 9.999620550574155e-06,
"loss": 0.4584,
"step": 264
},
{
"epoch": 15.647058823529411,
"grad_norm": 0.34673774242401123,
"learning_rate": 9.999433171682158e-06,
"loss": 0.3905,
"step": 266
},
{
"epoch": 15.764705882352942,
"grad_norm": 0.46017172932624817,
"learning_rate": 9.999208320101643e-06,
"loss": 0.4964,
"step": 268
},
{
"epoch": 15.882352941176471,
"grad_norm": 0.5312222838401794,
"learning_rate": 9.998945997517957e-06,
"loss": 0.3088,
"step": 270
},
{
"epoch": 16.0,
"grad_norm": 0.5483973026275635,
"learning_rate": 9.99864620589731e-06,
"loss": 0.4813,
"step": 272
},
{
"epoch": 16.0,
"eval_loss": 0.5214402079582214,
"eval_runtime": 31.8457,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 272
},
{
"epoch": 16.11764705882353,
"grad_norm": 0.40592437982559204,
"learning_rate": 9.998308947486753e-06,
"loss": 0.274,
"step": 274
},
{
"epoch": 16.235294117647058,
"grad_norm": 0.4894661009311676,
"learning_rate": 9.997934224814173e-06,
"loss": 0.361,
"step": 276
},
{
"epoch": 16.352941176470587,
"grad_norm": 0.5340806841850281,
"learning_rate": 9.997522040688258e-06,
"loss": 0.4403,
"step": 278
},
{
"epoch": 16.470588235294116,
"grad_norm": 0.6398447751998901,
"learning_rate": 9.997072398198492e-06,
"loss": 0.6416,
"step": 280
},
{
"epoch": 16.58823529411765,
"grad_norm": 0.4408299922943115,
"learning_rate": 9.996585300715117e-06,
"loss": 0.3814,
"step": 282
},
{
"epoch": 16.705882352941178,
"grad_norm": 0.45829370617866516,
"learning_rate": 9.996060751889114e-06,
"loss": 0.4494,
"step": 284
},
{
"epoch": 16.823529411764707,
"grad_norm": 0.5563424229621887,
"learning_rate": 9.995498755652186e-06,
"loss": 0.297,
"step": 286
},
{
"epoch": 16.941176470588236,
"grad_norm": 0.423210084438324,
"learning_rate": 9.994899316216709e-06,
"loss": 0.3015,
"step": 288
},
{
"epoch": 17.0,
"eval_loss": 0.511587381362915,
"eval_runtime": 31.8593,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 289
},
{
"epoch": 17.058823529411764,
"grad_norm": 0.46335744857788086,
"learning_rate": 9.994262438075713e-06,
"loss": 0.4301,
"step": 290
},
{
"epoch": 17.176470588235293,
"grad_norm": 0.5282275676727295,
"learning_rate": 9.993588126002848e-06,
"loss": 0.3083,
"step": 292
},
{
"epoch": 17.294117647058822,
"grad_norm": 0.6960487961769104,
"learning_rate": 9.992876385052346e-06,
"loss": 0.5187,
"step": 294
},
{
"epoch": 17.41176470588235,
"grad_norm": 0.47753918170928955,
"learning_rate": 9.992127220558976e-06,
"loss": 0.2761,
"step": 296
},
{
"epoch": 17.529411764705884,
"grad_norm": 0.49851253628730774,
"learning_rate": 9.991340638138022e-06,
"loss": 0.2777,
"step": 298
},
{
"epoch": 17.647058823529413,
"grad_norm": 0.5102773904800415,
"learning_rate": 9.990516643685222e-06,
"loss": 0.4131,
"step": 300
},
{
"epoch": 17.764705882352942,
"grad_norm": 0.47633931040763855,
"learning_rate": 9.98965524337673e-06,
"loss": 0.5281,
"step": 302
},
{
"epoch": 17.88235294117647,
"grad_norm": 0.33498552441596985,
"learning_rate": 9.988756443669081e-06,
"loss": 0.2696,
"step": 304
},
{
"epoch": 18.0,
"grad_norm": 0.5153418183326721,
"learning_rate": 9.987820251299121e-06,
"loss": 0.3513,
"step": 306
},
{
"epoch": 18.0,
"eval_loss": 0.5070950388908386,
"eval_runtime": 31.8477,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 306
},
{
"epoch": 18.11764705882353,
"grad_norm": 0.3366803824901581,
"learning_rate": 9.98684667328398e-06,
"loss": 0.1873,
"step": 308
},
{
"epoch": 18.235294117647058,
"grad_norm": 0.6154054403305054,
"learning_rate": 9.985835716921e-06,
"loss": 0.391,
"step": 310
},
{
"epoch": 18.352941176470587,
"grad_norm": 0.6151386499404907,
"learning_rate": 9.984787389787689e-06,
"loss": 0.5524,
"step": 312
},
{
"epoch": 18.470588235294116,
"grad_norm": 0.34717944264411926,
"learning_rate": 9.983701699741668e-06,
"loss": 0.214,
"step": 314
},
{
"epoch": 18.58823529411765,
"grad_norm": 0.5958353877067566,
"learning_rate": 9.982578654920601e-06,
"loss": 0.4503,
"step": 316
},
{
"epoch": 18.705882352941178,
"grad_norm": 0.45278429985046387,
"learning_rate": 9.981418263742148e-06,
"loss": 0.2803,
"step": 318
},
{
"epoch": 18.823529411764707,
"grad_norm": 1.5949257612228394,
"learning_rate": 9.980220534903889e-06,
"loss": 0.3574,
"step": 320
},
{
"epoch": 18.941176470588236,
"grad_norm": 0.6842171549797058,
"learning_rate": 9.978985477383264e-06,
"loss": 0.3638,
"step": 322
},
{
"epoch": 19.0,
"eval_loss": 0.5485844016075134,
"eval_runtime": 31.8529,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 323
},
{
"epoch": 19.058823529411764,
"grad_norm": 0.5384336113929749,
"learning_rate": 9.97771310043751e-06,
"loss": 0.4372,
"step": 324
},
{
"epoch": 19.176470588235293,
"grad_norm": 18.550317764282227,
"learning_rate": 9.97640341360358e-06,
"loss": 0.3407,
"step": 326
},
{
"epoch": 19.294117647058822,
"grad_norm": 2.0581817626953125,
"learning_rate": 9.975056426698094e-06,
"loss": 0.2623,
"step": 328
},
{
"epoch": 19.41176470588235,
"grad_norm": 0.6005405187606812,
"learning_rate": 9.973672149817232e-06,
"loss": 0.258,
"step": 330
},
{
"epoch": 19.529411764705884,
"grad_norm": 0.8219459056854248,
"learning_rate": 9.972250593336689e-06,
"loss": 0.3986,
"step": 332
},
{
"epoch": 19.647058823529413,
"grad_norm": 1.8439754247665405,
"learning_rate": 9.970791767911581e-06,
"loss": 0.2887,
"step": 334
},
{
"epoch": 19.764705882352942,
"grad_norm": 2.892220973968506,
"learning_rate": 9.96929568447637e-06,
"loss": 0.3433,
"step": 336
},
{
"epoch": 19.88235294117647,
"grad_norm": 0.6804599761962891,
"learning_rate": 9.967762354244778e-06,
"loss": 0.3686,
"step": 338
},
{
"epoch": 20.0,
"grad_norm": 0.8408982753753662,
"learning_rate": 9.966191788709716e-06,
"loss": 0.5246,
"step": 340
},
{
"epoch": 20.0,
"eval_loss": 0.48128727078437805,
"eval_runtime": 31.8515,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 340
},
{
"epoch": 20.11764705882353,
"grad_norm": 0.5616675019264221,
"learning_rate": 9.964583999643174e-06,
"loss": 0.2985,
"step": 342
},
{
"epoch": 20.235294117647058,
"grad_norm": 0.4580283463001251,
"learning_rate": 9.962938999096159e-06,
"loss": 0.2537,
"step": 344
},
{
"epoch": 20.352941176470587,
"grad_norm": 0.857369601726532,
"learning_rate": 9.961256799398584e-06,
"loss": 0.2804,
"step": 346
},
{
"epoch": 20.470588235294116,
"grad_norm": 1.1619380712509155,
"learning_rate": 9.95953741315919e-06,
"loss": 0.519,
"step": 348
},
{
"epoch": 20.58823529411765,
"grad_norm": 0.5293999314308167,
"learning_rate": 9.957780853265441e-06,
"loss": 0.2812,
"step": 350
},
{
"epoch": 20.705882352941178,
"grad_norm": 0.6322659850120544,
"learning_rate": 9.955987132883435e-06,
"loss": 0.2476,
"step": 352
},
{
"epoch": 20.823529411764707,
"grad_norm": 0.6491472125053406,
"learning_rate": 9.954156265457801e-06,
"loss": 0.2483,
"step": 354
},
{
"epoch": 20.941176470588236,
"grad_norm": 0.869428813457489,
"learning_rate": 9.952288264711601e-06,
"loss": 0.4751,
"step": 356
},
{
"epoch": 21.0,
"eval_loss": 0.536905825138092,
"eval_runtime": 31.8461,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 357
},
{
"epoch": 21.058823529411764,
"grad_norm": 0.7692728042602539,
"learning_rate": 9.950383144646221e-06,
"loss": 0.4331,
"step": 358
},
{
"epoch": 21.176470588235293,
"grad_norm": 0.6134036779403687,
"learning_rate": 9.948440919541277e-06,
"loss": 0.3302,
"step": 360
},
{
"epoch": 21.294117647058822,
"grad_norm": 0.8637880682945251,
"learning_rate": 9.946461603954499e-06,
"loss": 0.5403,
"step": 362
},
{
"epoch": 21.41176470588235,
"grad_norm": 0.7665501832962036,
"learning_rate": 9.944445212721619e-06,
"loss": 0.2682,
"step": 364
},
{
"epoch": 21.529411764705884,
"grad_norm": 0.6994332671165466,
"learning_rate": 9.942391760956277e-06,
"loss": 0.4364,
"step": 366
},
{
"epoch": 21.647058823529413,
"grad_norm": 0.8093287944793701,
"learning_rate": 9.940301264049885e-06,
"loss": 0.2887,
"step": 368
},
{
"epoch": 21.764705882352942,
"grad_norm": 0.7530141472816467,
"learning_rate": 9.938173737671531e-06,
"loss": 0.2479,
"step": 370
},
{
"epoch": 21.88235294117647,
"grad_norm": 0.35253098607063293,
"learning_rate": 9.936009197767847e-06,
"loss": 0.1859,
"step": 372
},
{
"epoch": 22.0,
"grad_norm": 0.8288949728012085,
"learning_rate": 9.933807660562898e-06,
"loss": 0.2074,
"step": 374
},
{
"epoch": 22.0,
"eval_loss": 0.5176519751548767,
"eval_runtime": 31.8467,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 374
},
{
"epoch": 22.11764705882353,
"grad_norm": 0.7651407122612,
"learning_rate": 9.931569142558057e-06,
"loss": 0.3298,
"step": 376
},
{
"epoch": 22.235294117647058,
"grad_norm": 0.7411403059959412,
"learning_rate": 9.929293660531889e-06,
"loss": 0.3866,
"step": 378
},
{
"epoch": 22.352941176470587,
"grad_norm": 0.8898152709007263,
"learning_rate": 9.926981231540007e-06,
"loss": 0.3071,
"step": 380
},
{
"epoch": 22.470588235294116,
"grad_norm": 1.235049843788147,
"learning_rate": 9.924631872914967e-06,
"loss": 0.3644,
"step": 382
},
{
"epoch": 22.58823529411765,
"grad_norm": 0.9050986766815186,
"learning_rate": 9.922245602266119e-06,
"loss": 0.2616,
"step": 384
},
{
"epoch": 22.705882352941178,
"grad_norm": 0.986419141292572,
"learning_rate": 9.919822437479488e-06,
"loss": 0.2003,
"step": 386
},
{
"epoch": 22.823529411764707,
"grad_norm": 1.1197755336761475,
"learning_rate": 9.91736239671763e-06,
"loss": 0.3103,
"step": 388
},
{
"epoch": 22.941176470588236,
"grad_norm": 1.1346862316131592,
"learning_rate": 9.91486549841951e-06,
"loss": 0.2513,
"step": 390
},
{
"epoch": 23.0,
"eval_loss": 0.5108680129051208,
"eval_runtime": 31.8499,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 391
},
{
"epoch": 23.058823529411764,
"grad_norm": 0.3815372884273529,
"learning_rate": 9.912331761300341e-06,
"loss": 0.1238,
"step": 392
},
{
"epoch": 23.176470588235293,
"grad_norm": 0.004847167991101742,
"learning_rate": 9.909761204351469e-06,
"loss": 0.0666,
"step": 394
},
{
"epoch": 23.294117647058822,
"grad_norm": 0.9460958242416382,
"learning_rate": 9.90715384684021e-06,
"loss": 0.4942,
"step": 396
},
{
"epoch": 23.41176470588235,
"grad_norm": 1.0769481658935547,
"learning_rate": 9.904509708309723e-06,
"loss": 0.2983,
"step": 398
},
{
"epoch": 23.529411764705884,
"grad_norm": 0.9719020128250122,
"learning_rate": 9.901828808578846e-06,
"loss": 0.2081,
"step": 400
},
{
"epoch": 23.647058823529413,
"grad_norm": 1.7906395196914673,
"learning_rate": 9.899111167741966e-06,
"loss": 0.3697,
"step": 402
},
{
"epoch": 23.764705882352942,
"grad_norm": 1.5768994092941284,
"learning_rate": 9.896356806168851e-06,
"loss": 0.3432,
"step": 404
},
{
"epoch": 23.88235294117647,
"grad_norm": 0.9857346415519714,
"learning_rate": 9.89356574450451e-06,
"loss": 0.203,
"step": 406
},
{
"epoch": 24.0,
"grad_norm": 1.2476699352264404,
"learning_rate": 9.890738003669029e-06,
"loss": 0.3019,
"step": 408
},
{
"epoch": 24.0,
"eval_loss": 0.5099675059318542,
"eval_runtime": 31.8826,
"eval_samples_per_second": 1.129,
"eval_steps_per_second": 1.129,
"step": 408
},
{
"epoch": 24.11764705882353,
"grad_norm": 1.1214749813079834,
"learning_rate": 9.887873604857424e-06,
"loss": 0.3208,
"step": 410
},
{
"epoch": 24.235294117647058,
"grad_norm": 1.2565882205963135,
"learning_rate": 9.884972569539471e-06,
"loss": 0.2897,
"step": 412
},
{
"epoch": 24.352941176470587,
"grad_norm": 0.9853368401527405,
"learning_rate": 9.882034919459556e-06,
"loss": 0.2316,
"step": 414
},
{
"epoch": 24.470588235294116,
"grad_norm": 1.3978081941604614,
"learning_rate": 9.879060676636502e-06,
"loss": 0.3282,
"step": 416
},
{
"epoch": 24.58823529411765,
"grad_norm": 1.6501458883285522,
"learning_rate": 9.876049863363415e-06,
"loss": 0.2489,
"step": 418
},
{
"epoch": 24.705882352941178,
"grad_norm": 1.0865875482559204,
"learning_rate": 9.873002502207502e-06,
"loss": 0.1472,
"step": 420
},
{
"epoch": 24.823529411764707,
"grad_norm": 1.1791576147079468,
"learning_rate": 9.86991861600992e-06,
"loss": 0.1987,
"step": 422
},
{
"epoch": 24.941176470588236,
"grad_norm": 1.4849443435668945,
"learning_rate": 9.866798227885588e-06,
"loss": 0.2039,
"step": 424
},
{
"epoch": 25.0,
"eval_loss": 0.542856752872467,
"eval_runtime": 31.8504,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 425
},
{
"epoch": 25.058823529411764,
"grad_norm": 1.064898133277893,
"learning_rate": 9.863641361223025e-06,
"loss": 0.2573,
"step": 426
},
{
"epoch": 25.176470588235293,
"grad_norm": 1.3044614791870117,
"learning_rate": 9.860448039684169e-06,
"loss": 0.2652,
"step": 428
},
{
"epoch": 25.294117647058822,
"grad_norm": 1.1536026000976562,
"learning_rate": 9.857218287204204e-06,
"loss": 0.1707,
"step": 430
},
{
"epoch": 25.41176470588235,
"grad_norm": 1.575131893157959,
"learning_rate": 9.853952127991374e-06,
"loss": 0.2226,
"step": 432
},
{
"epoch": 25.529411764705884,
"grad_norm": 1.8576124906539917,
"learning_rate": 9.850649586526808e-06,
"loss": 0.1639,
"step": 434
},
{
"epoch": 25.647058823529413,
"grad_norm": 1.8986772298812866,
"learning_rate": 9.847310687564335e-06,
"loss": 0.367,
"step": 436
},
{
"epoch": 25.764705882352942,
"grad_norm": 1.0811642408370972,
"learning_rate": 9.843935456130295e-06,
"loss": 0.2656,
"step": 438
},
{
"epoch": 25.88235294117647,
"grad_norm": 1.5638922452926636,
"learning_rate": 9.840523917523354e-06,
"loss": 0.1376,
"step": 440
},
{
"epoch": 26.0,
"grad_norm": 1.4766254425048828,
"learning_rate": 9.83707609731432e-06,
"loss": 0.228,
"step": 442
},
{
"epoch": 26.0,
"eval_loss": 0.5160675644874573,
"eval_runtime": 31.8614,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 442
},
{
"epoch": 26.11764705882353,
"grad_norm": 0.983343780040741,
"learning_rate": 9.833592021345938e-06,
"loss": 0.1742,
"step": 444
},
{
"epoch": 26.235294117647058,
"grad_norm": 1.3315422534942627,
"learning_rate": 9.830071715732708e-06,
"loss": 0.1853,
"step": 446
},
{
"epoch": 26.352941176470587,
"grad_norm": 0.6733668446540833,
"learning_rate": 9.826515206860683e-06,
"loss": 0.1493,
"step": 448
},
{
"epoch": 26.470588235294116,
"grad_norm": 1.4821362495422363,
"learning_rate": 9.822922521387277e-06,
"loss": 0.2319,
"step": 450
},
{
"epoch": 26.58823529411765,
"grad_norm": 3.0113203525543213,
"learning_rate": 9.819293686241057e-06,
"loss": 0.253,
"step": 452
},
{
"epoch": 26.705882352941178,
"grad_norm": 2.0522031784057617,
"learning_rate": 9.81562872862155e-06,
"loss": 0.1756,
"step": 454
},
{
"epoch": 26.823529411764707,
"grad_norm": 2.2314300537109375,
"learning_rate": 9.811927675999035e-06,
"loss": 0.1977,
"step": 456
},
{
"epoch": 26.941176470588236,
"grad_norm": 1.571890115737915,
"learning_rate": 9.808190556114333e-06,
"loss": 0.2127,
"step": 458
},
{
"epoch": 27.0,
"eval_loss": 0.5205699801445007,
"eval_runtime": 31.8543,
"eval_samples_per_second": 1.13,
"eval_steps_per_second": 1.13,
"step": 459
},
{
"epoch": 27.0,
"step": 459,
"total_flos": 2.8798901486092288e+17,
"train_loss": 0.8010637635823689,
"train_runtime": 10567.3661,
"train_samples_per_second": 1.93,
"train_steps_per_second": 0.241
}
],
"logging_steps": 2,
"max_steps": 2550,
"num_input_tokens_seen": 0,
"num_train_epochs": 150,
"save_steps": 25,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 7,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.8798901486092288e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}