CodingQueen13 commited on
Commit
e3427c3
1 Parent(s): b904edb

End of training

Browse files
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  license: apache-2.0
3
  base_model: ntu-spml/distilhubert
4
  tags:
@@ -22,7 +23,7 @@ model-index:
22
  metrics:
23
  - name: Accuracy
24
  type: accuracy
25
- value: 0.84
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,8 +33,8 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [ntu-spml/distilhubert](https://huggingface.co/ntu-spml/distilhubert) on the GTZAN dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.8931
36
- - Accuracy: 0.84
37
 
38
  ## Model description
39
 
@@ -58,39 +59,34 @@ The following hyperparameters were used during training:
58
  - seed: 42
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: cosine
61
- - lr_scheduler_warmup_ratio: 0.05
62
- - num_epochs: 20
63
  - mixed_precision_training: Native AMP
64
 
65
  ### Training results
66
 
67
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
68
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
69
- | 1.9292 | 1.0 | 113 | 1.8783 | 0.49 |
70
- | 1.1771 | 2.0 | 226 | 1.2910 | 0.64 |
71
- | 0.9343 | 3.0 | 339 | 1.0554 | 0.67 |
72
- | 0.5929 | 4.0 | 452 | 0.8423 | 0.75 |
73
- | 0.5678 | 5.0 | 565 | 0.7333 | 0.78 |
74
- | 0.2351 | 6.0 | 678 | 0.6266 | 0.84 |
75
- | 0.2348 | 7.0 | 791 | 0.6047 | 0.81 |
76
- | 0.0834 | 8.0 | 904 | 0.7525 | 0.82 |
77
- | 0.0374 | 9.0 | 1017 | 0.8083 | 0.82 |
78
- | 0.0089 | 10.0 | 1130 | 0.7745 | 0.83 |
79
- | 0.0068 | 11.0 | 1243 | 0.7883 | 0.85 |
80
- | 0.0046 | 12.0 | 1356 | 0.8266 | 0.84 |
81
- | 0.004 | 13.0 | 1469 | 0.8557 | 0.84 |
82
- | 0.0038 | 14.0 | 1582 | 0.9485 | 0.83 |
83
- | 0.0031 | 15.0 | 1695 | 0.9042 | 0.84 |
84
- | 0.0031 | 16.0 | 1808 | 0.8945 | 0.84 |
85
- | 0.0031 | 17.0 | 1921 | 0.8780 | 0.84 |
86
- | 0.0026 | 18.0 | 2034 | 0.9071 | 0.84 |
87
- | 0.0027 | 19.0 | 2147 | 0.8932 | 0.84 |
88
- | 0.0027 | 20.0 | 2260 | 0.8931 | 0.84 |
89
 
90
 
91
  ### Framework versions
92
 
93
- - Transformers 4.42.4
94
  - Pytorch 2.4.0+cu121
95
  - Datasets 2.21.0
96
  - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: ntu-spml/distilhubert
5
  tags:
 
23
  metrics:
24
  - name: Accuracy
25
  type: accuracy
26
+ value: 0.82
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
33
 
34
  This model is a fine-tuned version of [ntu-spml/distilhubert](https://huggingface.co/ntu-spml/distilhubert) on the GTZAN dataset.
35
  It achieves the following results on the evaluation set:
36
+ - Loss: 0.6191
37
+ - Accuracy: 0.82
38
 
39
  ## Model description
40
 
 
59
  - seed: 42
60
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
  - lr_scheduler_type: cosine
62
+ - lr_scheduler_warmup_ratio: 0.1
63
+ - num_epochs: 15
64
  - mixed_precision_training: Native AMP
65
 
66
  ### Training results
67
 
68
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
69
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
70
+ | 2.1554 | 1.0 | 113 | 2.0427 | 0.44 |
71
+ | 1.5528 | 2.0 | 226 | 1.5599 | 0.5 |
72
+ | 1.3212 | 3.0 | 339 | 1.1755 | 0.6 |
73
+ | 0.9075 | 4.0 | 452 | 0.9560 | 0.73 |
74
+ | 0.7823 | 5.0 | 565 | 0.8967 | 0.74 |
75
+ | 0.7262 | 6.0 | 678 | 0.6578 | 0.8 |
76
+ | 0.5761 | 7.0 | 791 | 0.6274 | 0.81 |
77
+ | 0.3797 | 8.0 | 904 | 0.6923 | 0.82 |
78
+ | 0.4168 | 9.0 | 1017 | 0.5700 | 0.84 |
79
+ | 0.2646 | 10.0 | 1130 | 0.6484 | 0.81 |
80
+ | 0.1952 | 11.0 | 1243 | 0.5925 | 0.84 |
81
+ | 0.1403 | 12.0 | 1356 | 0.6551 | 0.82 |
82
+ | 0.1558 | 13.0 | 1469 | 0.6271 | 0.82 |
83
+ | 0.4606 | 14.0 | 1582 | 0.6272 | 0.82 |
84
+ | 0.2095 | 15.0 | 1695 | 0.6191 | 0.82 |
 
 
 
 
 
85
 
86
 
87
  ### Framework versions
88
 
89
+ - Transformers 4.44.2
90
  - Pytorch 2.4.0+cu121
91
  - Datasets 2.21.0
92
  - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85c13090413ddf30e7ef23ccf424277e1aa4466e51a3d65457711bb84166dd2a
3
  size 94771728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:703a6ffbfc9fb1c04374eaac4a9ce3e7b2b6f354eb6a307622b58b8e4d36a6cf
3
  size 94771728
runs/Sep09_09-51-15_c3f9c4d4f413/events.out.tfevents.1725875477.c3f9c4d4f413.533.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:089feac9f5f879b359e920b2e32eefff3380c8c71d2ef4008805a942aa47e518
3
- size 82071
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afe61bcecadfa0c87fe928d9a34650c9f769a0a9c8fe6f15688a43e51880ed69
3
+ size 82748