TerryHenrickson commited on
Commit
8396b43
1 Parent(s): 44a7c95

End of training

Browse files
README.md CHANGED
@@ -17,12 +17,12 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.2397
21
- - Rouge1: 39.9145
22
- - Rouge2: 33.183
23
- - Rougel: 39.9484
24
- - Rougelsum: 39.9376
25
- - Gen Len: 19.0
26
 
27
  ## Model description
28
 
@@ -47,33 +47,53 @@ The following hyperparameters were used during training:
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
- - num_epochs: 20
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Training results
54
 
55
- | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
56
- |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
57
- | No log | 1.0 | 90 | 3.4654 | 9.8007 | 0.2231 | 9.4993 | 9.501 | 18.6889 |
58
- | No log | 2.0 | 180 | 2.1906 | 10.4443 | 0.2158 | 9.8828 | 9.8755 | 18.8972 |
59
- | No log | 3.0 | 270 | 1.3067 | 11.7213 | 0.5145 | 10.913 | 10.9345 | 18.8528 |
60
- | No log | 4.0 | 360 | 0.7369 | 14.9807 | 1.4227 | 13.7179 | 13.7131 | 18.8333 |
61
- | No log | 5.0 | 450 | 0.6143 | 19.8511 | 4.6089 | 18.0244 | 17.9558 | 18.7083 |
62
- | 2.447 | 6.0 | 540 | 0.5312 | 23.1026 | 8.6515 | 20.7866 | 20.757 | 18.7139 |
63
- | 2.447 | 7.0 | 630 | 0.4782 | 21.9961 | 9.3626 | 19.7651 | 19.7488 | 18.5944 |
64
- | 2.447 | 8.0 | 720 | 0.4365 | 16.4406 | 6.5397 | 14.9694 | 14.9816 | 18.6639 |
65
- | 2.447 | 9.0 | 810 | 0.3603 | 6.9337 | 3.7397 | 6.6337 | 6.6621 | 18.9028 |
66
- | 2.447 | 10.0 | 900 | 0.2696 | 24.2884 | 19.0601 | 24.1044 | 24.1488 | 18.9694 |
67
- | 2.447 | 11.0 | 990 | 0.2590 | 39.2002 | 32.3107 | 39.202 | 39.1928 | 19.0 |
68
- | 0.572 | 12.0 | 1080 | 0.2546 | 39.0083 | 32.1464 | 39.0296 | 38.9988 | 19.0 |
69
- | 0.572 | 13.0 | 1170 | 0.2486 | 39.519 | 32.7114 | 39.5614 | 39.5391 | 19.0 |
70
- | 0.572 | 14.0 | 1260 | 0.2465 | 39.589 | 32.8014 | 39.6298 | 39.6092 | 19.0 |
71
- | 0.572 | 15.0 | 1350 | 0.2444 | 39.5831 | 32.7959 | 39.6266 | 39.6123 | 19.0 |
72
- | 0.572 | 16.0 | 1440 | 0.2427 | 39.7174 | 32.9525 | 39.7513 | 39.7311 | 19.0 |
73
- | 0.3469 | 17.0 | 1530 | 0.2412 | 39.8478 | 33.0999 | 39.8901 | 39.8708 | 19.0 |
74
- | 0.3469 | 18.0 | 1620 | 0.2401 | 39.8528 | 33.1031 | 39.8819 | 39.873 | 19.0 |
75
- | 0.3469 | 19.0 | 1710 | 0.2398 | 39.9283 | 33.1964 | 39.9502 | 39.9533 | 19.0 |
76
- | 0.3469 | 20.0 | 1800 | 0.2397 | 39.9145 | 33.183 | 39.9484 | 39.9376 | 19.0 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77
 
78
 
79
  ### Framework versions
 
17
 
18
  This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.0734
21
+ - Rouge1: 99.9038
22
+ - Rouge2: 99.838
23
+ - Rougel: 99.9145
24
+ - Rougelsum: 99.9038
25
+ - Gen Len: 93.9181
26
 
27
  ## Model description
28
 
 
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
+ - num_epochs: 40
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Training results
54
 
55
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
56
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:--------:|
57
+ | No log | 1.0 | 180 | 1.7815 | 9.7268 | 2.7047 | 8.7069 | 8.7035 | 155.8472 |
58
+ | No log | 2.0 | 360 | 0.6270 | 28.7135 | 19.99 | 27.1646 | 27.1386 | 265.2903 |
59
+ | 2.122 | 3.0 | 540 | 0.3572 | 21.4211 | 17.5143 | 21.0387 | 20.9118 | 142.7333 |
60
+ | 2.122 | 4.0 | 720 | 0.2757 | 92.8223 | 90.5077 | 92.0061 | 92.0015 | 87.0847 |
61
+ | 2.122 | 5.0 | 900 | 0.2493 | 95.6972 | 94.5082 | 95.5057 | 95.522 | 91.8556 |
62
+ | 0.4002 | 6.0 | 1080 | 0.2348 | 96.8942 | 96.2704 | 96.7552 | 96.7736 | 96.0764 |
63
+ | 0.4002 | 7.0 | 1260 | 0.2227 | 97.7669 | 97.4255 | 97.6867 | 97.6913 | 93.9097 |
64
+ | 0.4002 | 8.0 | 1440 | 0.2111 | 98.7823 | 98.5538 | 98.7622 | 98.7722 | 94.2875 |
65
+ | 0.2717 | 9.0 | 1620 | 0.1979 | 99.7455 | 99.6524 | 99.7428 | 99.7449 | 93.8569 |
66
+ | 0.2717 | 10.0 | 1800 | 0.1843 | 99.8967 | 99.8175 | 99.8953 | 99.8939 | 93.875 |
67
+ | 0.2717 | 11.0 | 1980 | 0.1716 | 99.9078 | 99.8578 | 99.9114 | 99.9095 | 93.8556 |
68
+ | 0.2244 | 12.0 | 2160 | 0.1606 | 99.9371 | 99.8807 | 99.9373 | 99.9373 | 93.9236 |
69
+ | 0.2244 | 13.0 | 2340 | 0.1512 | 99.9112 | 99.8535 | 99.9141 | 99.9103 | 93.8542 |
70
+ | 0.19 | 14.0 | 2520 | 0.1424 | 99.9573 | 99.919 | 99.9573 | 99.9573 | 93.9236 |
71
+ | 0.19 | 15.0 | 2700 | 0.1353 | 99.9679 | 99.9421 | 99.9679 | 99.9679 | 93.925 |
72
+ | 0.19 | 16.0 | 2880 | 0.1290 | 99.9234 | 99.8727 | 99.9323 | 99.9234 | 93.8736 |
73
+ | 0.1652 | 17.0 | 3060 | 0.1235 | 99.9252 | 99.8727 | 99.9359 | 99.9252 | 93.9222 |
74
+ | 0.1652 | 18.0 | 3240 | 0.1184 | 99.9038 | 99.8373 | 99.911 | 99.9021 | 93.8722 |
75
+ | 0.1652 | 19.0 | 3420 | 0.1137 | 99.9466 | 99.9074 | 99.9573 | 99.9466 | 93.9236 |
76
+ | 0.1471 | 20.0 | 3600 | 0.1092 | 99.9252 | 99.8727 | 99.9359 | 99.9252 | 93.9222 |
77
+ | 0.1471 | 21.0 | 3780 | 0.1053 | 99.9252 | 99.8727 | 99.9359 | 99.9252 | 93.9222 |
78
+ | 0.1471 | 22.0 | 3960 | 0.1014 | 99.9252 | 99.8727 | 99.9359 | 99.9252 | 93.9222 |
79
+ | 0.1331 | 23.0 | 4140 | 0.0982 | 99.9252 | 99.8727 | 99.9359 | 99.9252 | 93.9222 |
80
+ | 0.1331 | 24.0 | 4320 | 0.0949 | 99.9252 | 99.8727 | 99.9359 | 99.9252 | 93.9208 |
81
+ | 0.1226 | 25.0 | 4500 | 0.0918 | 99.9252 | 99.8727 | 99.9359 | 99.9252 | 93.9208 |
82
+ | 0.1226 | 26.0 | 4680 | 0.0892 | 99.9252 | 99.8727 | 99.9359 | 99.9252 | 93.9208 |
83
+ | 0.1226 | 27.0 | 4860 | 0.0867 | 99.9252 | 99.8727 | 99.9359 | 99.9252 | 93.9208 |
84
+ | 0.114 | 28.0 | 5040 | 0.0848 | 99.9145 | 99.8495 | 99.9252 | 99.9145 | 93.9194 |
85
+ | 0.114 | 29.0 | 5220 | 0.0828 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
86
+ | 0.114 | 30.0 | 5400 | 0.0811 | 99.9145 | 99.8495 | 99.9252 | 99.9145 | 93.9194 |
87
+ | 0.1074 | 31.0 | 5580 | 0.0794 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
88
+ | 0.1074 | 32.0 | 5760 | 0.0781 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
89
+ | 0.1074 | 33.0 | 5940 | 0.0769 | 99.9252 | 99.8669 | 99.9252 | 99.9252 | 93.9194 |
90
+ | 0.1027 | 34.0 | 6120 | 0.0757 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
91
+ | 0.1027 | 35.0 | 6300 | 0.0751 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
92
+ | 0.1027 | 36.0 | 6480 | 0.0745 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
93
+ | 0.0994 | 37.0 | 6660 | 0.0740 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
94
+ | 0.0994 | 38.0 | 6840 | 0.0737 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
95
+ | 0.0975 | 39.0 | 7020 | 0.0735 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
96
+ | 0.0975 | 40.0 | 7200 | 0.0734 | 99.9038 | 99.838 | 99.9145 | 99.9038 | 93.9181 |
97
 
98
 
99
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:10c4cb2d5e25b466a3451ccebc6704e9dc799ae4018c6b62282efd2023533c50
3
  size 241990696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cc7a243653ddad1ab21f7a5da717592a544b7c2777704ba5f31a8e5800d2d83
3
  size 241990696
runs/Feb19_09-07-25_3e139e734b2e/events.out.tfevents.1708333646.3e139e734b2e.332.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8dfae86c54986efdf131e17de81bafad438ba34954efd75d65d515efb3a38177
3
- size 27361
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8654accfaa1b9f903bfc78ce9c5d40671e7e22f1bb9ca61cef3d7fb5fa9e556c
3
+ size 28765