LN-Thai-14B-v0.1 / trainer_state.json
CjangCjengh's picture
Upload files
a6053dc
raw
history blame
34.4 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9970879440885265,
"eval_steps": 500,
"global_step": 214,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004659289458357601,
"grad_norm": 37.4553865599607,
"learning_rate": 0.0,
"loss": 2.4488,
"step": 1
},
{
"epoch": 0.009318578916715201,
"grad_norm": 10.88300608738222,
"learning_rate": 6.309297535714574e-06,
"loss": 2.1388,
"step": 2
},
{
"epoch": 0.013977868375072802,
"grad_norm": 11.568553101687147,
"learning_rate": 1e-05,
"loss": 2.2252,
"step": 3
},
{
"epoch": 0.018637157833430402,
"grad_norm": 7.643720965332644,
"learning_rate": 1e-05,
"loss": 1.9329,
"step": 4
},
{
"epoch": 0.023296447291788,
"grad_norm": 5.748382614323999,
"learning_rate": 1e-05,
"loss": 1.8794,
"step": 5
},
{
"epoch": 0.027955736750145604,
"grad_norm": 6.567036028997142,
"learning_rate": 1e-05,
"loss": 1.8839,
"step": 6
},
{
"epoch": 0.032615026208503206,
"grad_norm": 4.4306758077305455,
"learning_rate": 1e-05,
"loss": 1.8774,
"step": 7
},
{
"epoch": 0.037274315666860805,
"grad_norm": 3.660970589217281,
"learning_rate": 1e-05,
"loss": 1.7406,
"step": 8
},
{
"epoch": 0.041933605125218404,
"grad_norm": 3.2695948007985227,
"learning_rate": 1e-05,
"loss": 1.8113,
"step": 9
},
{
"epoch": 0.046592894583576,
"grad_norm": 3.356468736985637,
"learning_rate": 1e-05,
"loss": 1.7207,
"step": 10
},
{
"epoch": 0.0512521840419336,
"grad_norm": 3.5450444591858394,
"learning_rate": 1e-05,
"loss": 1.7032,
"step": 11
},
{
"epoch": 0.05591147350029121,
"grad_norm": 3.19419501302636,
"learning_rate": 1e-05,
"loss": 1.699,
"step": 12
},
{
"epoch": 0.060570762958648806,
"grad_norm": 2.8474267589247417,
"learning_rate": 1e-05,
"loss": 1.6978,
"step": 13
},
{
"epoch": 0.06523005241700641,
"grad_norm": 3.1135350288269166,
"learning_rate": 1e-05,
"loss": 1.7062,
"step": 14
},
{
"epoch": 0.06988934187536401,
"grad_norm": 2.770146289905479,
"learning_rate": 1e-05,
"loss": 1.6356,
"step": 15
},
{
"epoch": 0.07454863133372161,
"grad_norm": 2.9479050208283173,
"learning_rate": 1e-05,
"loss": 1.6605,
"step": 16
},
{
"epoch": 0.07920792079207921,
"grad_norm": 2.9102020117907847,
"learning_rate": 1e-05,
"loss": 1.6725,
"step": 17
},
{
"epoch": 0.08386721025043681,
"grad_norm": 2.837755280153835,
"learning_rate": 1e-05,
"loss": 1.7021,
"step": 18
},
{
"epoch": 0.0885264997087944,
"grad_norm": 2.7877405064893557,
"learning_rate": 1e-05,
"loss": 1.5947,
"step": 19
},
{
"epoch": 0.093185789167152,
"grad_norm": 2.5815355221611727,
"learning_rate": 1e-05,
"loss": 1.6915,
"step": 20
},
{
"epoch": 0.0978450786255096,
"grad_norm": 2.7006002228913526,
"learning_rate": 1e-05,
"loss": 1.6286,
"step": 21
},
{
"epoch": 0.1025043680838672,
"grad_norm": 2.784209069755651,
"learning_rate": 1e-05,
"loss": 1.6685,
"step": 22
},
{
"epoch": 0.10716365754222482,
"grad_norm": 2.9524143186766567,
"learning_rate": 1e-05,
"loss": 1.6315,
"step": 23
},
{
"epoch": 0.11182294700058241,
"grad_norm": 3.070122490149911,
"learning_rate": 1e-05,
"loss": 1.647,
"step": 24
},
{
"epoch": 0.11648223645894001,
"grad_norm": 2.719822294385443,
"learning_rate": 1e-05,
"loss": 1.6034,
"step": 25
},
{
"epoch": 0.12114152591729761,
"grad_norm": 2.925248468241921,
"learning_rate": 1e-05,
"loss": 1.6777,
"step": 26
},
{
"epoch": 0.1258008153756552,
"grad_norm": 2.9031875873598962,
"learning_rate": 1e-05,
"loss": 1.63,
"step": 27
},
{
"epoch": 0.13046010483401282,
"grad_norm": 2.6836639962275184,
"learning_rate": 1e-05,
"loss": 1.638,
"step": 28
},
{
"epoch": 0.1351193942923704,
"grad_norm": 2.5540688972245515,
"learning_rate": 1e-05,
"loss": 1.6721,
"step": 29
},
{
"epoch": 0.13977868375072802,
"grad_norm": 2.811652255435364,
"learning_rate": 1e-05,
"loss": 1.602,
"step": 30
},
{
"epoch": 0.1444379732090856,
"grad_norm": 2.570946004882331,
"learning_rate": 1e-05,
"loss": 1.6118,
"step": 31
},
{
"epoch": 0.14909726266744322,
"grad_norm": 2.371184921508139,
"learning_rate": 1e-05,
"loss": 1.6038,
"step": 32
},
{
"epoch": 0.1537565521258008,
"grad_norm": 2.677393543132455,
"learning_rate": 1e-05,
"loss": 1.6211,
"step": 33
},
{
"epoch": 0.15841584158415842,
"grad_norm": 2.8813028493341495,
"learning_rate": 1e-05,
"loss": 1.6526,
"step": 34
},
{
"epoch": 0.163075131042516,
"grad_norm": 2.517387020615115,
"learning_rate": 1e-05,
"loss": 1.5879,
"step": 35
},
{
"epoch": 0.16773442050087362,
"grad_norm": 2.3301459869544527,
"learning_rate": 1e-05,
"loss": 1.5954,
"step": 36
},
{
"epoch": 0.17239370995923123,
"grad_norm": 2.4447934986055335,
"learning_rate": 1e-05,
"loss": 1.5929,
"step": 37
},
{
"epoch": 0.1770529994175888,
"grad_norm": 2.817763614899898,
"learning_rate": 1e-05,
"loss": 1.6128,
"step": 38
},
{
"epoch": 0.18171228887594643,
"grad_norm": 2.566684690956894,
"learning_rate": 1e-05,
"loss": 1.5668,
"step": 39
},
{
"epoch": 0.186371578334304,
"grad_norm": 2.661562404310152,
"learning_rate": 1e-05,
"loss": 1.6154,
"step": 40
},
{
"epoch": 0.19103086779266162,
"grad_norm": 2.4627832136882444,
"learning_rate": 1e-05,
"loss": 1.6519,
"step": 41
},
{
"epoch": 0.1956901572510192,
"grad_norm": 2.597954766071769,
"learning_rate": 1e-05,
"loss": 1.6,
"step": 42
},
{
"epoch": 0.20034944670937682,
"grad_norm": 2.754493319360813,
"learning_rate": 1e-05,
"loss": 1.5652,
"step": 43
},
{
"epoch": 0.2050087361677344,
"grad_norm": 2.6151045585662045,
"learning_rate": 1e-05,
"loss": 1.5709,
"step": 44
},
{
"epoch": 0.20966802562609202,
"grad_norm": 2.394985883263223,
"learning_rate": 1e-05,
"loss": 1.5708,
"step": 45
},
{
"epoch": 0.21432731508444963,
"grad_norm": 2.754166542274207,
"learning_rate": 1e-05,
"loss": 1.5709,
"step": 46
},
{
"epoch": 0.21898660454280722,
"grad_norm": 2.363803590882057,
"learning_rate": 1e-05,
"loss": 1.5652,
"step": 47
},
{
"epoch": 0.22364589400116483,
"grad_norm": 2.779877793574347,
"learning_rate": 1e-05,
"loss": 1.5415,
"step": 48
},
{
"epoch": 0.22830518345952241,
"grad_norm": 2.6675042398192534,
"learning_rate": 1e-05,
"loss": 1.6183,
"step": 49
},
{
"epoch": 0.23296447291788003,
"grad_norm": 2.4116922018773095,
"learning_rate": 1e-05,
"loss": 1.5527,
"step": 50
},
{
"epoch": 0.2376237623762376,
"grad_norm": 2.3315932179612155,
"learning_rate": 1e-05,
"loss": 1.5901,
"step": 51
},
{
"epoch": 0.24228305183459523,
"grad_norm": 2.3019835184053905,
"learning_rate": 1e-05,
"loss": 1.544,
"step": 52
},
{
"epoch": 0.24694234129295284,
"grad_norm": 2.6424581659593898,
"learning_rate": 1e-05,
"loss": 1.5892,
"step": 53
},
{
"epoch": 0.2516016307513104,
"grad_norm": 2.587045419024974,
"learning_rate": 1e-05,
"loss": 1.577,
"step": 54
},
{
"epoch": 0.256260920209668,
"grad_norm": 2.4773093313847685,
"learning_rate": 1e-05,
"loss": 1.5864,
"step": 55
},
{
"epoch": 0.26092020966802565,
"grad_norm": 2.4149084409040116,
"learning_rate": 1e-05,
"loss": 1.6176,
"step": 56
},
{
"epoch": 0.26557949912638323,
"grad_norm": 2.4512017044947734,
"learning_rate": 1e-05,
"loss": 1.6246,
"step": 57
},
{
"epoch": 0.2702387885847408,
"grad_norm": 2.401483628381168,
"learning_rate": 1e-05,
"loss": 1.5652,
"step": 58
},
{
"epoch": 0.2748980780430984,
"grad_norm": 2.5214336145047,
"learning_rate": 1e-05,
"loss": 1.5791,
"step": 59
},
{
"epoch": 0.27955736750145604,
"grad_norm": 3.024295336906611,
"learning_rate": 1e-05,
"loss": 1.5421,
"step": 60
},
{
"epoch": 0.28421665695981363,
"grad_norm": 2.509650308333821,
"learning_rate": 1e-05,
"loss": 1.5147,
"step": 61
},
{
"epoch": 0.2888759464181712,
"grad_norm": 2.374198285202519,
"learning_rate": 1e-05,
"loss": 1.5759,
"step": 62
},
{
"epoch": 0.29353523587652885,
"grad_norm": 2.3112309742555066,
"learning_rate": 1e-05,
"loss": 1.5253,
"step": 63
},
{
"epoch": 0.29819452533488644,
"grad_norm": 2.3977382823518303,
"learning_rate": 1e-05,
"loss": 1.5422,
"step": 64
},
{
"epoch": 0.302853814793244,
"grad_norm": 2.508604177665278,
"learning_rate": 1e-05,
"loss": 1.5628,
"step": 65
},
{
"epoch": 0.3075131042516016,
"grad_norm": 2.5543152862332685,
"learning_rate": 1e-05,
"loss": 1.5789,
"step": 66
},
{
"epoch": 0.31217239370995925,
"grad_norm": 2.615684261823402,
"learning_rate": 1e-05,
"loss": 1.497,
"step": 67
},
{
"epoch": 0.31683168316831684,
"grad_norm": 2.321913113942261,
"learning_rate": 1e-05,
"loss": 1.5619,
"step": 68
},
{
"epoch": 0.3214909726266744,
"grad_norm": 2.3380073471748783,
"learning_rate": 1e-05,
"loss": 1.5434,
"step": 69
},
{
"epoch": 0.326150262085032,
"grad_norm": 2.194756331960441,
"learning_rate": 1e-05,
"loss": 1.5429,
"step": 70
},
{
"epoch": 0.33080955154338965,
"grad_norm": 2.339565902888178,
"learning_rate": 1e-05,
"loss": 1.5898,
"step": 71
},
{
"epoch": 0.33546884100174723,
"grad_norm": 2.589829563994277,
"learning_rate": 1e-05,
"loss": 1.522,
"step": 72
},
{
"epoch": 0.3401281304601048,
"grad_norm": 2.475665557332239,
"learning_rate": 1e-05,
"loss": 1.6017,
"step": 73
},
{
"epoch": 0.34478741991846246,
"grad_norm": 2.638214948295715,
"learning_rate": 1e-05,
"loss": 1.4913,
"step": 74
},
{
"epoch": 0.34944670937682004,
"grad_norm": 2.373158176130504,
"learning_rate": 1e-05,
"loss": 1.5273,
"step": 75
},
{
"epoch": 0.3541059988351776,
"grad_norm": 2.3444375276771323,
"learning_rate": 1e-05,
"loss": 1.5152,
"step": 76
},
{
"epoch": 0.3587652882935352,
"grad_norm": 2.2407053882148142,
"learning_rate": 1e-05,
"loss": 1.5944,
"step": 77
},
{
"epoch": 0.36342457775189285,
"grad_norm": 2.337831790434085,
"learning_rate": 1e-05,
"loss": 1.547,
"step": 78
},
{
"epoch": 0.36808386721025044,
"grad_norm": 2.5112280582608864,
"learning_rate": 1e-05,
"loss": 1.5366,
"step": 79
},
{
"epoch": 0.372743156668608,
"grad_norm": 2.258790618824264,
"learning_rate": 1e-05,
"loss": 1.5493,
"step": 80
},
{
"epoch": 0.37740244612696566,
"grad_norm": 2.255161437661444,
"learning_rate": 1e-05,
"loss": 1.5308,
"step": 81
},
{
"epoch": 0.38206173558532325,
"grad_norm": 2.4688158937749485,
"learning_rate": 1e-05,
"loss": 1.5546,
"step": 82
},
{
"epoch": 0.38672102504368083,
"grad_norm": 2.4069475693862867,
"learning_rate": 1e-05,
"loss": 1.5211,
"step": 83
},
{
"epoch": 0.3913803145020384,
"grad_norm": 2.374794651480313,
"learning_rate": 1e-05,
"loss": 1.494,
"step": 84
},
{
"epoch": 0.39603960396039606,
"grad_norm": 2.3398152209467833,
"learning_rate": 1e-05,
"loss": 1.5342,
"step": 85
},
{
"epoch": 0.40069889341875364,
"grad_norm": 2.35428983704996,
"learning_rate": 1e-05,
"loss": 1.4785,
"step": 86
},
{
"epoch": 0.40535818287711123,
"grad_norm": 2.388724058110063,
"learning_rate": 1e-05,
"loss": 1.5467,
"step": 87
},
{
"epoch": 0.4100174723354688,
"grad_norm": 2.4134839455509804,
"learning_rate": 1e-05,
"loss": 1.5179,
"step": 88
},
{
"epoch": 0.41467676179382645,
"grad_norm": 2.463768285922605,
"learning_rate": 1e-05,
"loss": 1.576,
"step": 89
},
{
"epoch": 0.41933605125218404,
"grad_norm": 2.444581005858159,
"learning_rate": 1e-05,
"loss": 1.5008,
"step": 90
},
{
"epoch": 0.4239953407105416,
"grad_norm": 2.190503733939241,
"learning_rate": 1e-05,
"loss": 1.5156,
"step": 91
},
{
"epoch": 0.42865463016889926,
"grad_norm": 2.593833508710404,
"learning_rate": 1e-05,
"loss": 1.5075,
"step": 92
},
{
"epoch": 0.43331391962725685,
"grad_norm": 2.2851452541574218,
"learning_rate": 1e-05,
"loss": 1.5325,
"step": 93
},
{
"epoch": 0.43797320908561443,
"grad_norm": 2.3210289097461776,
"learning_rate": 1e-05,
"loss": 1.5011,
"step": 94
},
{
"epoch": 0.442632498543972,
"grad_norm": 2.3246717326280044,
"learning_rate": 1e-05,
"loss": 1.5085,
"step": 95
},
{
"epoch": 0.44729178800232966,
"grad_norm": 2.4021880355407546,
"learning_rate": 1e-05,
"loss": 1.5572,
"step": 96
},
{
"epoch": 0.45195107746068724,
"grad_norm": 2.1612576748589825,
"learning_rate": 1e-05,
"loss": 1.5181,
"step": 97
},
{
"epoch": 0.45661036691904483,
"grad_norm": 2.517060089072893,
"learning_rate": 1e-05,
"loss": 1.5696,
"step": 98
},
{
"epoch": 0.46126965637740247,
"grad_norm": 2.509779517463924,
"learning_rate": 1e-05,
"loss": 1.4737,
"step": 99
},
{
"epoch": 0.46592894583576006,
"grad_norm": 2.2278628463607597,
"learning_rate": 1e-05,
"loss": 1.4381,
"step": 100
},
{
"epoch": 0.47058823529411764,
"grad_norm": 2.5421872957086853,
"learning_rate": 1e-05,
"loss": 1.5216,
"step": 101
},
{
"epoch": 0.4752475247524752,
"grad_norm": 2.7534020196807623,
"learning_rate": 1e-05,
"loss": 1.5311,
"step": 102
},
{
"epoch": 0.47990681421083287,
"grad_norm": 2.315666594456174,
"learning_rate": 1e-05,
"loss": 1.5471,
"step": 103
},
{
"epoch": 0.48456610366919045,
"grad_norm": 2.181276723760835,
"learning_rate": 1e-05,
"loss": 1.5186,
"step": 104
},
{
"epoch": 0.48922539312754804,
"grad_norm": 2.3599566582977856,
"learning_rate": 1e-05,
"loss": 1.5002,
"step": 105
},
{
"epoch": 0.4938846825859057,
"grad_norm": 2.4800646557316255,
"learning_rate": 1e-05,
"loss": 1.4866,
"step": 106
},
{
"epoch": 0.49854397204426326,
"grad_norm": 2.7615250141518275,
"learning_rate": 1e-05,
"loss": 1.4466,
"step": 107
},
{
"epoch": 0.5032032615026208,
"grad_norm": 2.1541119559811968,
"learning_rate": 1e-05,
"loss": 1.4961,
"step": 108
},
{
"epoch": 0.5078625509609784,
"grad_norm": 2.496900445760294,
"learning_rate": 1e-05,
"loss": 1.5489,
"step": 109
},
{
"epoch": 0.512521840419336,
"grad_norm": 2.3790745422703923,
"learning_rate": 1e-05,
"loss": 1.4719,
"step": 110
},
{
"epoch": 0.5171811298776936,
"grad_norm": 2.5485805451241963,
"learning_rate": 1e-05,
"loss": 1.4548,
"step": 111
},
{
"epoch": 0.5218404193360513,
"grad_norm": 2.5177273826789257,
"learning_rate": 1e-05,
"loss": 1.5397,
"step": 112
},
{
"epoch": 0.5264997087944089,
"grad_norm": 2.1272963564434737,
"learning_rate": 1e-05,
"loss": 1.4996,
"step": 113
},
{
"epoch": 0.5311589982527665,
"grad_norm": 2.50068908262842,
"learning_rate": 1e-05,
"loss": 1.558,
"step": 114
},
{
"epoch": 0.535818287711124,
"grad_norm": 2.214031615505479,
"learning_rate": 1e-05,
"loss": 1.5344,
"step": 115
},
{
"epoch": 0.5404775771694816,
"grad_norm": 2.399654826141851,
"learning_rate": 1e-05,
"loss": 1.476,
"step": 116
},
{
"epoch": 0.5451368666278392,
"grad_norm": 2.593360531751168,
"learning_rate": 1e-05,
"loss": 1.4059,
"step": 117
},
{
"epoch": 0.5497961560861968,
"grad_norm": 2.2733077198422302,
"learning_rate": 1e-05,
"loss": 1.5092,
"step": 118
},
{
"epoch": 0.5544554455445545,
"grad_norm": 2.6231459607903087,
"learning_rate": 1e-05,
"loss": 1.5377,
"step": 119
},
{
"epoch": 0.5591147350029121,
"grad_norm": 3.2448905784760154,
"learning_rate": 1e-05,
"loss": 1.4897,
"step": 120
},
{
"epoch": 0.5637740244612697,
"grad_norm": 2.695228082395806,
"learning_rate": 1e-05,
"loss": 1.4672,
"step": 121
},
{
"epoch": 0.5684333139196273,
"grad_norm": 2.3198621060308517,
"learning_rate": 1e-05,
"loss": 1.4536,
"step": 122
},
{
"epoch": 0.5730926033779848,
"grad_norm": 2.424456337054707,
"learning_rate": 1e-05,
"loss": 1.501,
"step": 123
},
{
"epoch": 0.5777518928363424,
"grad_norm": 2.4598395558661865,
"learning_rate": 1e-05,
"loss": 1.5316,
"step": 124
},
{
"epoch": 0.5824111822947,
"grad_norm": 2.3377997659212277,
"learning_rate": 1e-05,
"loss": 1.506,
"step": 125
},
{
"epoch": 0.5870704717530577,
"grad_norm": 2.3201349850077664,
"learning_rate": 1e-05,
"loss": 1.5269,
"step": 126
},
{
"epoch": 0.5917297612114153,
"grad_norm": 2.414429631758563,
"learning_rate": 1e-05,
"loss": 1.5293,
"step": 127
},
{
"epoch": 0.5963890506697729,
"grad_norm": 2.416825959879323,
"learning_rate": 1e-05,
"loss": 1.4832,
"step": 128
},
{
"epoch": 0.6010483401281305,
"grad_norm": 2.597816910489374,
"learning_rate": 1e-05,
"loss": 1.4724,
"step": 129
},
{
"epoch": 0.605707629586488,
"grad_norm": 2.4148507856855925,
"learning_rate": 1e-05,
"loss": 1.4891,
"step": 130
},
{
"epoch": 0.6103669190448456,
"grad_norm": 2.210119122851114,
"learning_rate": 1e-05,
"loss": 1.577,
"step": 131
},
{
"epoch": 0.6150262085032032,
"grad_norm": 2.0971029486323456,
"learning_rate": 1e-05,
"loss": 1.5182,
"step": 132
},
{
"epoch": 0.6196854979615609,
"grad_norm": 2.4223761752873125,
"learning_rate": 1e-05,
"loss": 1.502,
"step": 133
},
{
"epoch": 0.6243447874199185,
"grad_norm": 2.260148223947814,
"learning_rate": 1e-05,
"loss": 1.5131,
"step": 134
},
{
"epoch": 0.6290040768782761,
"grad_norm": 2.2187749396883607,
"learning_rate": 1e-05,
"loss": 1.481,
"step": 135
},
{
"epoch": 0.6336633663366337,
"grad_norm": 2.3857038661729866,
"learning_rate": 1e-05,
"loss": 1.4215,
"step": 136
},
{
"epoch": 0.6383226557949913,
"grad_norm": 2.2285653535516454,
"learning_rate": 1e-05,
"loss": 1.4648,
"step": 137
},
{
"epoch": 0.6429819452533488,
"grad_norm": 2.2814076725542716,
"learning_rate": 1e-05,
"loss": 1.4983,
"step": 138
},
{
"epoch": 0.6476412347117064,
"grad_norm": 2.391062816273672,
"learning_rate": 1e-05,
"loss": 1.4481,
"step": 139
},
{
"epoch": 0.652300524170064,
"grad_norm": 2.172767929285484,
"learning_rate": 1e-05,
"loss": 1.5194,
"step": 140
},
{
"epoch": 0.6569598136284217,
"grad_norm": 2.985146157885572,
"learning_rate": 1e-05,
"loss": 1.425,
"step": 141
},
{
"epoch": 0.6616191030867793,
"grad_norm": 2.3548178966971784,
"learning_rate": 1e-05,
"loss": 1.4631,
"step": 142
},
{
"epoch": 0.6662783925451369,
"grad_norm": 2.328614489701982,
"learning_rate": 1e-05,
"loss": 1.5093,
"step": 143
},
{
"epoch": 0.6709376820034945,
"grad_norm": 2.3820404107843616,
"learning_rate": 1e-05,
"loss": 1.5127,
"step": 144
},
{
"epoch": 0.675596971461852,
"grad_norm": 2.4654257970703273,
"learning_rate": 1e-05,
"loss": 1.5163,
"step": 145
},
{
"epoch": 0.6802562609202096,
"grad_norm": 2.306376065137465,
"learning_rate": 1e-05,
"loss": 1.4831,
"step": 146
},
{
"epoch": 0.6849155503785672,
"grad_norm": 2.436381432968586,
"learning_rate": 1e-05,
"loss": 1.4877,
"step": 147
},
{
"epoch": 0.6895748398369249,
"grad_norm": 2.3572365071763524,
"learning_rate": 1e-05,
"loss": 1.5309,
"step": 148
},
{
"epoch": 0.6942341292952825,
"grad_norm": 2.4470550124596238,
"learning_rate": 1e-05,
"loss": 1.4608,
"step": 149
},
{
"epoch": 0.6988934187536401,
"grad_norm": 2.184405357759209,
"learning_rate": 1e-05,
"loss": 1.5079,
"step": 150
},
{
"epoch": 0.7035527082119977,
"grad_norm": 2.2624803921660392,
"learning_rate": 1e-05,
"loss": 1.506,
"step": 151
},
{
"epoch": 0.7082119976703553,
"grad_norm": 2.3185716811531303,
"learning_rate": 1e-05,
"loss": 1.465,
"step": 152
},
{
"epoch": 0.7128712871287128,
"grad_norm": 2.481541020027683,
"learning_rate": 1e-05,
"loss": 1.4744,
"step": 153
},
{
"epoch": 0.7175305765870704,
"grad_norm": 5.450013277465346,
"learning_rate": 1e-05,
"loss": 1.542,
"step": 154
},
{
"epoch": 0.7221898660454281,
"grad_norm": 2.488065611187432,
"learning_rate": 1e-05,
"loss": 1.4258,
"step": 155
},
{
"epoch": 0.7268491555037857,
"grad_norm": 2.6254843382308857,
"learning_rate": 1e-05,
"loss": 1.5095,
"step": 156
},
{
"epoch": 0.7315084449621433,
"grad_norm": 2.283524625320461,
"learning_rate": 1e-05,
"loss": 1.4876,
"step": 157
},
{
"epoch": 0.7361677344205009,
"grad_norm": 2.306519172530066,
"learning_rate": 1e-05,
"loss": 1.4782,
"step": 158
},
{
"epoch": 0.7408270238788585,
"grad_norm": 2.2310587388915946,
"learning_rate": 1e-05,
"loss": 1.4803,
"step": 159
},
{
"epoch": 0.745486313337216,
"grad_norm": 2.4218732296938468,
"learning_rate": 1e-05,
"loss": 1.5058,
"step": 160
},
{
"epoch": 0.7501456027955736,
"grad_norm": 2.2702106423890354,
"learning_rate": 1e-05,
"loss": 1.5014,
"step": 161
},
{
"epoch": 0.7548048922539313,
"grad_norm": 2.2465894583609405,
"learning_rate": 1e-05,
"loss": 1.4762,
"step": 162
},
{
"epoch": 0.7594641817122889,
"grad_norm": 2.3237613014258742,
"learning_rate": 1e-05,
"loss": 1.4943,
"step": 163
},
{
"epoch": 0.7641234711706465,
"grad_norm": 2.1190413494106246,
"learning_rate": 1e-05,
"loss": 1.5,
"step": 164
},
{
"epoch": 0.7687827606290041,
"grad_norm": 2.4483459352892223,
"learning_rate": 1e-05,
"loss": 1.519,
"step": 165
},
{
"epoch": 0.7734420500873617,
"grad_norm": 2.1538492135049263,
"learning_rate": 1e-05,
"loss": 1.4622,
"step": 166
},
{
"epoch": 0.7781013395457193,
"grad_norm": 2.2203603693455403,
"learning_rate": 1e-05,
"loss": 1.4886,
"step": 167
},
{
"epoch": 0.7827606290040768,
"grad_norm": 2.3648740483376964,
"learning_rate": 1e-05,
"loss": 1.5106,
"step": 168
},
{
"epoch": 0.7874199184624345,
"grad_norm": 2.3767712572720514,
"learning_rate": 1e-05,
"loss": 1.4777,
"step": 169
},
{
"epoch": 0.7920792079207921,
"grad_norm": 2.2620536568087526,
"learning_rate": 1e-05,
"loss": 1.4844,
"step": 170
},
{
"epoch": 0.7967384973791497,
"grad_norm": 2.136482480053942,
"learning_rate": 1e-05,
"loss": 1.5182,
"step": 171
},
{
"epoch": 0.8013977868375073,
"grad_norm": 2.258498047531436,
"learning_rate": 1e-05,
"loss": 1.4992,
"step": 172
},
{
"epoch": 0.8060570762958649,
"grad_norm": 2.2792395779463734,
"learning_rate": 1e-05,
"loss": 1.4773,
"step": 173
},
{
"epoch": 0.8107163657542225,
"grad_norm": 2.1881766755914396,
"learning_rate": 1e-05,
"loss": 1.4938,
"step": 174
},
{
"epoch": 0.81537565521258,
"grad_norm": 2.1353292001431536,
"learning_rate": 1e-05,
"loss": 1.4345,
"step": 175
},
{
"epoch": 0.8200349446709376,
"grad_norm": 2.2630090798181492,
"learning_rate": 1e-05,
"loss": 1.4636,
"step": 176
},
{
"epoch": 0.8246942341292953,
"grad_norm": 2.2293749498374127,
"learning_rate": 1e-05,
"loss": 1.432,
"step": 177
},
{
"epoch": 0.8293535235876529,
"grad_norm": 2.3221095280470023,
"learning_rate": 1e-05,
"loss": 1.4521,
"step": 178
},
{
"epoch": 0.8340128130460105,
"grad_norm": 2.385465927102841,
"learning_rate": 1e-05,
"loss": 1.5021,
"step": 179
},
{
"epoch": 0.8386721025043681,
"grad_norm": 2.2944255924050276,
"learning_rate": 1e-05,
"loss": 1.4978,
"step": 180
},
{
"epoch": 0.8433313919627257,
"grad_norm": 2.3098860073018956,
"learning_rate": 1e-05,
"loss": 1.5041,
"step": 181
},
{
"epoch": 0.8479906814210832,
"grad_norm": 2.3335048971904775,
"learning_rate": 1e-05,
"loss": 1.4943,
"step": 182
},
{
"epoch": 0.8526499708794408,
"grad_norm": 2.223659498237971,
"learning_rate": 1e-05,
"loss": 1.5338,
"step": 183
},
{
"epoch": 0.8573092603377985,
"grad_norm": 2.3293622100025058,
"learning_rate": 1e-05,
"loss": 1.4855,
"step": 184
},
{
"epoch": 0.8619685497961561,
"grad_norm": 2.408437081867758,
"learning_rate": 1e-05,
"loss": 1.4348,
"step": 185
},
{
"epoch": 0.8666278392545137,
"grad_norm": 2.289339067399062,
"learning_rate": 1e-05,
"loss": 1.4403,
"step": 186
},
{
"epoch": 0.8712871287128713,
"grad_norm": 2.2759064288074806,
"learning_rate": 1e-05,
"loss": 1.3777,
"step": 187
},
{
"epoch": 0.8759464181712289,
"grad_norm": 2.48715528854699,
"learning_rate": 1e-05,
"loss": 1.4431,
"step": 188
},
{
"epoch": 0.8806057076295865,
"grad_norm": 2.400786521041563,
"learning_rate": 1e-05,
"loss": 1.5096,
"step": 189
},
{
"epoch": 0.885264997087944,
"grad_norm": 2.2444464605037786,
"learning_rate": 1e-05,
"loss": 1.5071,
"step": 190
},
{
"epoch": 0.8899242865463017,
"grad_norm": 2.3631736662246556,
"learning_rate": 1e-05,
"loss": 1.5074,
"step": 191
},
{
"epoch": 0.8945835760046593,
"grad_norm": 2.270308066018393,
"learning_rate": 1e-05,
"loss": 1.4219,
"step": 192
},
{
"epoch": 0.8992428654630169,
"grad_norm": 2.1490689617434615,
"learning_rate": 1e-05,
"loss": 1.4453,
"step": 193
},
{
"epoch": 0.9039021549213745,
"grad_norm": 2.234379186727459,
"learning_rate": 1e-05,
"loss": 1.5035,
"step": 194
},
{
"epoch": 0.9085614443797321,
"grad_norm": 2.378879658628609,
"learning_rate": 1e-05,
"loss": 1.5105,
"step": 195
},
{
"epoch": 0.9132207338380897,
"grad_norm": 2.2408872339352586,
"learning_rate": 1e-05,
"loss": 1.5111,
"step": 196
},
{
"epoch": 0.9178800232964472,
"grad_norm": 2.3800770297089513,
"learning_rate": 1e-05,
"loss": 1.4406,
"step": 197
},
{
"epoch": 0.9225393127548049,
"grad_norm": 2.2789001631581143,
"learning_rate": 1e-05,
"loss": 1.4671,
"step": 198
},
{
"epoch": 0.9271986022131625,
"grad_norm": 2.3783253348876845,
"learning_rate": 1e-05,
"loss": 1.4054,
"step": 199
},
{
"epoch": 0.9318578916715201,
"grad_norm": 2.0882291082381372,
"learning_rate": 1e-05,
"loss": 1.4656,
"step": 200
},
{
"epoch": 0.9365171811298777,
"grad_norm": 2.3231028485545537,
"learning_rate": 1e-05,
"loss": 1.4347,
"step": 201
},
{
"epoch": 0.9411764705882353,
"grad_norm": 2.251299349081433,
"learning_rate": 1e-05,
"loss": 1.5202,
"step": 202
},
{
"epoch": 0.9458357600465929,
"grad_norm": 2.1273492280549466,
"learning_rate": 1e-05,
"loss": 1.511,
"step": 203
},
{
"epoch": 0.9504950495049505,
"grad_norm": 2.2583829949862078,
"learning_rate": 1e-05,
"loss": 1.472,
"step": 204
},
{
"epoch": 0.9551543389633081,
"grad_norm": 2.3080136541764817,
"learning_rate": 1e-05,
"loss": 1.4602,
"step": 205
},
{
"epoch": 0.9598136284216657,
"grad_norm": 2.2472965019702884,
"learning_rate": 1e-05,
"loss": 1.4436,
"step": 206
},
{
"epoch": 0.9644729178800233,
"grad_norm": 2.0993853167635415,
"learning_rate": 1e-05,
"loss": 1.538,
"step": 207
},
{
"epoch": 0.9691322073383809,
"grad_norm": 2.2063602589571536,
"learning_rate": 1e-05,
"loss": 1.459,
"step": 208
},
{
"epoch": 0.9737914967967385,
"grad_norm": 2.4073762011028457,
"learning_rate": 1e-05,
"loss": 1.4872,
"step": 209
},
{
"epoch": 0.9784507862550961,
"grad_norm": 2.2657217549149706,
"learning_rate": 1e-05,
"loss": 1.433,
"step": 210
},
{
"epoch": 0.9831100757134537,
"grad_norm": 2.250620730742056,
"learning_rate": 1e-05,
"loss": 1.4799,
"step": 211
},
{
"epoch": 0.9877693651718114,
"grad_norm": 2.32373073377042,
"learning_rate": 1e-05,
"loss": 1.5087,
"step": 212
},
{
"epoch": 0.9924286546301689,
"grad_norm": 2.2674019100510665,
"learning_rate": 1e-05,
"loss": 1.4293,
"step": 213
},
{
"epoch": 0.9970879440885265,
"grad_norm": 2.1053994055125633,
"learning_rate": 1e-05,
"loss": 1.4257,
"step": 214
},
{
"epoch": 0.9970879440885265,
"step": 214,
"total_flos": 346293314519040.0,
"train_loss": 1.5478776877171525,
"train_runtime": 30341.6487,
"train_samples_per_second": 0.905,
"train_steps_per_second": 0.007
}
],
"logging_steps": 1.0,
"max_steps": 214,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 10,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 346293314519040.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}