gemma-story-generator-finetuned / trainer_state.json
SarwarShafee's picture
Upload 11 files
85d0e0b verified
raw
history blame
14.8 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.3315579227696404,
"eval_steps": 500,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01,
"grad_norm": 1.4126338958740234,
"learning_rate": 0.0002,
"loss": 2.411,
"step": 10
},
{
"epoch": 0.03,
"grad_norm": 1.0683205127716064,
"learning_rate": 0.0002,
"loss": 1.7825,
"step": 20
},
{
"epoch": 0.04,
"grad_norm": 1.250962734222412,
"learning_rate": 0.0002,
"loss": 1.6118,
"step": 30
},
{
"epoch": 0.05,
"grad_norm": 1.128470778465271,
"learning_rate": 0.0002,
"loss": 1.4993,
"step": 40
},
{
"epoch": 0.07,
"grad_norm": 1.0976437330245972,
"learning_rate": 0.0002,
"loss": 1.538,
"step": 50
},
{
"epoch": 0.08,
"grad_norm": 1.0831102132797241,
"learning_rate": 0.0002,
"loss": 1.4524,
"step": 60
},
{
"epoch": 0.09,
"grad_norm": 1.1627174615859985,
"learning_rate": 0.0002,
"loss": 1.4061,
"step": 70
},
{
"epoch": 0.11,
"grad_norm": 1.1374784708023071,
"learning_rate": 0.0002,
"loss": 1.4085,
"step": 80
},
{
"epoch": 0.12,
"grad_norm": 1.2638696432113647,
"learning_rate": 0.0002,
"loss": 1.3816,
"step": 90
},
{
"epoch": 0.13,
"grad_norm": 1.0406217575073242,
"learning_rate": 0.0002,
"loss": 1.3639,
"step": 100
},
{
"epoch": 0.15,
"grad_norm": 1.1120456457138062,
"learning_rate": 0.0002,
"loss": 1.3145,
"step": 110
},
{
"epoch": 0.16,
"grad_norm": 1.0537450313568115,
"learning_rate": 0.0002,
"loss": 1.3204,
"step": 120
},
{
"epoch": 0.17,
"grad_norm": 1.0962114334106445,
"learning_rate": 0.0002,
"loss": 1.3467,
"step": 130
},
{
"epoch": 0.19,
"grad_norm": 1.0847195386886597,
"learning_rate": 0.0002,
"loss": 1.3216,
"step": 140
},
{
"epoch": 0.2,
"grad_norm": 1.1323330402374268,
"learning_rate": 0.0002,
"loss": 1.3546,
"step": 150
},
{
"epoch": 0.21,
"grad_norm": 1.1049960851669312,
"learning_rate": 0.0002,
"loss": 1.3617,
"step": 160
},
{
"epoch": 0.23,
"grad_norm": 1.0923413038253784,
"learning_rate": 0.0002,
"loss": 1.3248,
"step": 170
},
{
"epoch": 0.24,
"grad_norm": 1.128686547279358,
"learning_rate": 0.0002,
"loss": 1.3002,
"step": 180
},
{
"epoch": 0.25,
"grad_norm": 1.0518907308578491,
"learning_rate": 0.0002,
"loss": 1.3624,
"step": 190
},
{
"epoch": 0.27,
"grad_norm": 1.060835599899292,
"learning_rate": 0.0002,
"loss": 1.368,
"step": 200
},
{
"epoch": 0.28,
"grad_norm": 1.1328442096710205,
"learning_rate": 0.0002,
"loss": 1.2611,
"step": 210
},
{
"epoch": 0.29,
"grad_norm": 1.0428252220153809,
"learning_rate": 0.0002,
"loss": 1.2889,
"step": 220
},
{
"epoch": 0.31,
"grad_norm": 1.0666313171386719,
"learning_rate": 0.0002,
"loss": 1.2383,
"step": 230
},
{
"epoch": 0.32,
"grad_norm": 1.0708130598068237,
"learning_rate": 0.0002,
"loss": 1.2666,
"step": 240
},
{
"epoch": 0.33,
"grad_norm": 1.0374865531921387,
"learning_rate": 0.0002,
"loss": 1.231,
"step": 250
},
{
"epoch": 0.35,
"grad_norm": 1.02108895778656,
"learning_rate": 0.0002,
"loss": 1.2519,
"step": 260
},
{
"epoch": 0.36,
"grad_norm": 0.9618756175041199,
"learning_rate": 0.0002,
"loss": 1.2115,
"step": 270
},
{
"epoch": 0.37,
"grad_norm": 1.0125309228897095,
"learning_rate": 0.0002,
"loss": 1.1766,
"step": 280
},
{
"epoch": 0.39,
"grad_norm": 1.031528115272522,
"learning_rate": 0.0002,
"loss": 1.2598,
"step": 290
},
{
"epoch": 0.4,
"grad_norm": 1.1020361185073853,
"learning_rate": 0.0002,
"loss": 1.1946,
"step": 300
},
{
"epoch": 0.41,
"grad_norm": 1.0392624139785767,
"learning_rate": 0.0002,
"loss": 1.226,
"step": 310
},
{
"epoch": 0.43,
"grad_norm": 1.0464197397232056,
"learning_rate": 0.0002,
"loss": 1.2406,
"step": 320
},
{
"epoch": 0.44,
"grad_norm": 1.035586953163147,
"learning_rate": 0.0002,
"loss": 1.1949,
"step": 330
},
{
"epoch": 0.45,
"grad_norm": 1.053875207901001,
"learning_rate": 0.0002,
"loss": 1.2137,
"step": 340
},
{
"epoch": 0.47,
"grad_norm": 1.013602375984192,
"learning_rate": 0.0002,
"loss": 1.2162,
"step": 350
},
{
"epoch": 0.48,
"grad_norm": 1.0841283798217773,
"learning_rate": 0.0002,
"loss": 1.1973,
"step": 360
},
{
"epoch": 0.49,
"grad_norm": 0.9872801303863525,
"learning_rate": 0.0002,
"loss": 1.2096,
"step": 370
},
{
"epoch": 0.51,
"grad_norm": 0.9933327436447144,
"learning_rate": 0.0002,
"loss": 1.2049,
"step": 380
},
{
"epoch": 0.52,
"grad_norm": 1.0115890502929688,
"learning_rate": 0.0002,
"loss": 1.231,
"step": 390
},
{
"epoch": 0.53,
"grad_norm": 1.0036671161651611,
"learning_rate": 0.0002,
"loss": 1.225,
"step": 400
},
{
"epoch": 0.55,
"grad_norm": 1.027393102645874,
"learning_rate": 0.0002,
"loss": 1.1855,
"step": 410
},
{
"epoch": 0.56,
"grad_norm": 1.0328060388565063,
"learning_rate": 0.0002,
"loss": 1.1978,
"step": 420
},
{
"epoch": 0.57,
"grad_norm": 0.9716214537620544,
"learning_rate": 0.0002,
"loss": 1.2326,
"step": 430
},
{
"epoch": 0.59,
"grad_norm": 0.9971627593040466,
"learning_rate": 0.0002,
"loss": 1.2019,
"step": 440
},
{
"epoch": 0.6,
"grad_norm": 1.002035140991211,
"learning_rate": 0.0002,
"loss": 1.1845,
"step": 450
},
{
"epoch": 0.61,
"grad_norm": 1.0371512174606323,
"learning_rate": 0.0002,
"loss": 1.1668,
"step": 460
},
{
"epoch": 0.63,
"grad_norm": 1.0316438674926758,
"learning_rate": 0.0002,
"loss": 1.1967,
"step": 470
},
{
"epoch": 0.64,
"grad_norm": 1.0340334177017212,
"learning_rate": 0.0002,
"loss": 1.1564,
"step": 480
},
{
"epoch": 0.65,
"grad_norm": 1.0055785179138184,
"learning_rate": 0.0002,
"loss": 1.1749,
"step": 490
},
{
"epoch": 0.67,
"grad_norm": 0.9474852085113525,
"learning_rate": 0.0002,
"loss": 1.1847,
"step": 500
},
{
"epoch": 0.68,
"grad_norm": 0.979642927646637,
"learning_rate": 0.0002,
"loss": 1.213,
"step": 510
},
{
"epoch": 0.69,
"grad_norm": 1.0399181842803955,
"learning_rate": 0.0002,
"loss": 1.1641,
"step": 520
},
{
"epoch": 0.71,
"grad_norm": 0.9442302584648132,
"learning_rate": 0.0002,
"loss": 1.2084,
"step": 530
},
{
"epoch": 0.72,
"grad_norm": 1.038650393486023,
"learning_rate": 0.0002,
"loss": 1.1934,
"step": 540
},
{
"epoch": 0.73,
"grad_norm": 1.0110139846801758,
"learning_rate": 0.0002,
"loss": 1.1741,
"step": 550
},
{
"epoch": 0.75,
"grad_norm": 1.0062506198883057,
"learning_rate": 0.0002,
"loss": 1.2044,
"step": 560
},
{
"epoch": 0.76,
"grad_norm": 1.0056486129760742,
"learning_rate": 0.0002,
"loss": 1.1683,
"step": 570
},
{
"epoch": 0.77,
"grad_norm": 0.9406468272209167,
"learning_rate": 0.0002,
"loss": 1.2082,
"step": 580
},
{
"epoch": 0.79,
"grad_norm": 0.9766492247581482,
"learning_rate": 0.0002,
"loss": 1.1402,
"step": 590
},
{
"epoch": 0.8,
"grad_norm": 1.011161208152771,
"learning_rate": 0.0002,
"loss": 1.1836,
"step": 600
},
{
"epoch": 0.81,
"grad_norm": 0.9706726670265198,
"learning_rate": 0.0002,
"loss": 1.1364,
"step": 610
},
{
"epoch": 0.83,
"grad_norm": 1.060975193977356,
"learning_rate": 0.0002,
"loss": 1.1492,
"step": 620
},
{
"epoch": 0.84,
"grad_norm": 1.0223751068115234,
"learning_rate": 0.0002,
"loss": 1.1389,
"step": 630
},
{
"epoch": 0.85,
"grad_norm": 1.132886290550232,
"learning_rate": 0.0002,
"loss": 1.195,
"step": 640
},
{
"epoch": 0.87,
"grad_norm": 0.9614752531051636,
"learning_rate": 0.0002,
"loss": 1.1469,
"step": 650
},
{
"epoch": 0.88,
"grad_norm": 1.0038574934005737,
"learning_rate": 0.0002,
"loss": 1.1901,
"step": 660
},
{
"epoch": 0.89,
"grad_norm": 1.0092958211898804,
"learning_rate": 0.0002,
"loss": 1.1215,
"step": 670
},
{
"epoch": 0.91,
"grad_norm": 1.0625172853469849,
"learning_rate": 0.0002,
"loss": 1.1552,
"step": 680
},
{
"epoch": 0.92,
"grad_norm": 0.9498062133789062,
"learning_rate": 0.0002,
"loss": 1.1454,
"step": 690
},
{
"epoch": 0.93,
"grad_norm": 0.9746269583702087,
"learning_rate": 0.0002,
"loss": 1.1322,
"step": 700
},
{
"epoch": 0.95,
"grad_norm": 0.9876968860626221,
"learning_rate": 0.0002,
"loss": 1.1473,
"step": 710
},
{
"epoch": 0.96,
"grad_norm": 0.9368664622306824,
"learning_rate": 0.0002,
"loss": 1.1036,
"step": 720
},
{
"epoch": 0.97,
"grad_norm": 0.9563239216804504,
"learning_rate": 0.0002,
"loss": 1.124,
"step": 730
},
{
"epoch": 0.99,
"grad_norm": 0.9494954347610474,
"learning_rate": 0.0002,
"loss": 1.0844,
"step": 740
},
{
"epoch": 1.0,
"grad_norm": 0.9336316585540771,
"learning_rate": 0.0002,
"loss": 1.0938,
"step": 750
},
{
"epoch": 1.01,
"grad_norm": 0.9765089154243469,
"learning_rate": 0.0002,
"loss": 1.0245,
"step": 760
},
{
"epoch": 1.03,
"grad_norm": 1.0190706253051758,
"learning_rate": 0.0002,
"loss": 1.0672,
"step": 770
},
{
"epoch": 1.04,
"grad_norm": 1.0273311138153076,
"learning_rate": 0.0002,
"loss": 1.0371,
"step": 780
},
{
"epoch": 1.05,
"grad_norm": 1.0353707075119019,
"learning_rate": 0.0002,
"loss": 1.0212,
"step": 790
},
{
"epoch": 1.07,
"grad_norm": 0.9515432119369507,
"learning_rate": 0.0002,
"loss": 1.0721,
"step": 800
},
{
"epoch": 1.08,
"grad_norm": 1.0564075708389282,
"learning_rate": 0.0002,
"loss": 1.0495,
"step": 810
},
{
"epoch": 1.09,
"grad_norm": 1.004117727279663,
"learning_rate": 0.0002,
"loss": 1.032,
"step": 820
},
{
"epoch": 1.11,
"grad_norm": 1.0284982919692993,
"learning_rate": 0.0002,
"loss": 1.0621,
"step": 830
},
{
"epoch": 1.12,
"grad_norm": 0.9780373573303223,
"learning_rate": 0.0002,
"loss": 1.0241,
"step": 840
},
{
"epoch": 1.13,
"grad_norm": 0.9741361141204834,
"learning_rate": 0.0002,
"loss": 1.0932,
"step": 850
},
{
"epoch": 1.15,
"grad_norm": 1.0497852563858032,
"learning_rate": 0.0002,
"loss": 1.0665,
"step": 860
},
{
"epoch": 1.16,
"grad_norm": 1.0611335039138794,
"learning_rate": 0.0002,
"loss": 1.0485,
"step": 870
},
{
"epoch": 1.17,
"grad_norm": 1.0571497678756714,
"learning_rate": 0.0002,
"loss": 1.0452,
"step": 880
},
{
"epoch": 1.19,
"grad_norm": 1.038238525390625,
"learning_rate": 0.0002,
"loss": 1.0088,
"step": 890
},
{
"epoch": 1.2,
"grad_norm": 1.0717825889587402,
"learning_rate": 0.0002,
"loss": 1.0708,
"step": 900
},
{
"epoch": 1.21,
"grad_norm": 1.1264833211898804,
"learning_rate": 0.0002,
"loss": 1.0645,
"step": 910
},
{
"epoch": 1.23,
"grad_norm": 1.0441728830337524,
"learning_rate": 0.0002,
"loss": 1.0637,
"step": 920
},
{
"epoch": 1.24,
"grad_norm": 1.0234287977218628,
"learning_rate": 0.0002,
"loss": 1.0639,
"step": 930
},
{
"epoch": 1.25,
"grad_norm": 1.043578028678894,
"learning_rate": 0.0002,
"loss": 1.0185,
"step": 940
},
{
"epoch": 1.26,
"grad_norm": 1.0319379568099976,
"learning_rate": 0.0002,
"loss": 1.035,
"step": 950
},
{
"epoch": 1.28,
"grad_norm": 1.0528068542480469,
"learning_rate": 0.0002,
"loss": 1.0313,
"step": 960
},
{
"epoch": 1.29,
"grad_norm": 0.9674711227416992,
"learning_rate": 0.0002,
"loss": 1.051,
"step": 970
},
{
"epoch": 1.3,
"grad_norm": 1.0464591979980469,
"learning_rate": 0.0002,
"loss": 1.0588,
"step": 980
},
{
"epoch": 1.32,
"grad_norm": 1.0295021533966064,
"learning_rate": 0.0002,
"loss": 1.0356,
"step": 990
},
{
"epoch": 1.33,
"grad_norm": 1.0435758829116821,
"learning_rate": 0.0002,
"loss": 1.0621,
"step": 1000
}
],
"logging_steps": 10,
"max_steps": 1000,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 10,
"total_flos": 9.789555867648e+16,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}