haryoaw commited on
Commit
69476e4
1 Parent(s): bd7acc6

Initial Commit

Browse files
Files changed (4) hide show
  1. README.md +27 -16
  2. config.json +21 -14
  3. pytorch_model.bin +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: mit
3
- base_model: facebook/xlm-v-base
4
  tags:
5
  - generated_from_trainer
6
  datasets:
@@ -23,10 +23,10 @@ model-index:
23
  metrics:
24
  - name: Accuracy
25
  type: accuracy
26
- value: 0.051647811116576486
27
  - name: F1
28
  type: f1
29
- value: 0.0016647904742274576
30
  ---
31
 
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -34,11 +34,11 @@ should probably proofread and complete it, then remove this comment. -->
34
 
35
  # scenario-TCR_data-AmazonScience_massive_all_1_1
36
 
37
- This model is a fine-tuned version of [facebook/xlm-v-base](https://huggingface.co/facebook/xlm-v-base) on the massive dataset.
38
  It achieves the following results on the evaluation set:
39
- - Loss: 3.8189
40
- - Accuracy: 0.0516
41
- - F1: 0.0017
42
 
43
  ## Model description
44
 
@@ -60,22 +60,33 @@ The following hyperparameters were used during training:
60
  - learning_rate: 5e-05
61
  - train_batch_size: 32
62
  - eval_batch_size: 32
63
- - seed: 42
64
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
  - lr_scheduler_type: linear
66
- - num_epochs: 500
67
 
68
  ### Training results
69
 
70
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
71
  |:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|
72
- | 3.7541 | 0.27 | 5000 | 3.7437 | 0.0605 | 0.0019 |
73
- | 3.7191 | 0.53 | 10000 | 3.7655 | 0.0620 | 0.0020 |
74
- | 3.7281 | 0.8 | 15000 | 3.8011 | 0.0516 | 0.0017 |
75
- | 3.7116 | 1.07 | 20000 | 3.8592 | 0.0516 | 0.0017 |
76
- | 3.7068 | 1.34 | 25000 | 3.8168 | 0.0516 | 0.0017 |
77
- | 3.7064 | 1.6 | 30000 | 3.8746 | 0.0516 | 0.0017 |
78
- | 3.7059 | 1.87 | 35000 | 3.8189 | 0.0516 | 0.0017 |
 
 
 
 
 
 
 
 
 
 
 
79
 
80
 
81
  ### Framework versions
 
1
  ---
2
  license: mit
3
+ base_model: microsoft/mdeberta-v3-base
4
  tags:
5
  - generated_from_trainer
6
  datasets:
 
23
  metrics:
24
  - name: Accuracy
25
  type: accuracy
26
+ value: 0.8558780127889818
27
  - name: F1
28
  type: f1
29
+ value: 0.8318635435156069
30
  ---
31
 
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
34
 
35
  # scenario-TCR_data-AmazonScience_massive_all_1_1
36
 
37
+ This model is a fine-tuned version of [microsoft/mdeberta-v3-base](https://huggingface.co/microsoft/mdeberta-v3-base) on the massive dataset.
38
  It achieves the following results on the evaluation set:
39
+ - Loss: 0.9483
40
+ - Accuracy: 0.8559
41
+ - F1: 0.8319
42
 
43
  ## Model description
44
 
 
60
  - learning_rate: 5e-05
61
  - train_batch_size: 32
62
  - eval_batch_size: 32
63
+ - seed: 66
64
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
  - lr_scheduler_type: linear
66
+ - num_epochs: 30
67
 
68
  ### Training results
69
 
70
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
71
  |:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|
72
+ | 0.519 | 0.27 | 5000 | 0.6915 | 0.8379 | 0.7941 |
73
+ | 0.3806 | 0.53 | 10000 | 0.6969 | 0.8468 | 0.8063 |
74
+ | 0.3259 | 0.8 | 15000 | 0.6916 | 0.8515 | 0.8159 |
75
+ | 0.2379 | 1.07 | 20000 | 0.7826 | 0.8505 | 0.8191 |
76
+ | 0.236 | 1.34 | 25000 | 0.7514 | 0.8508 | 0.8189 |
77
+ | 0.2298 | 1.6 | 30000 | 0.7719 | 0.8526 | 0.8267 |
78
+ | 0.2169 | 1.87 | 35000 | 0.8162 | 0.8505 | 0.8265 |
79
+ | 0.164 | 2.14 | 40000 | 0.8316 | 0.8549 | 0.8272 |
80
+ | 0.1684 | 2.41 | 45000 | 0.8123 | 0.8513 | 0.8204 |
81
+ | 0.158 | 2.67 | 50000 | 0.8252 | 0.8556 | 0.8309 |
82
+ | 0.1761 | 2.94 | 55000 | 0.8092 | 0.8545 | 0.8287 |
83
+ | 0.1378 | 3.21 | 60000 | 0.8574 | 0.8607 | 0.8357 |
84
+ | 0.1399 | 3.47 | 65000 | 0.8976 | 0.8572 | 0.8359 |
85
+ | 0.1431 | 3.74 | 70000 | 0.8908 | 0.8536 | 0.8350 |
86
+ | 0.1249 | 4.01 | 75000 | 0.9613 | 0.8533 | 0.8292 |
87
+ | 0.1129 | 4.28 | 80000 | 0.9511 | 0.8543 | 0.8306 |
88
+ | 0.1143 | 4.54 | 85000 | 0.9001 | 0.8563 | 0.8331 |
89
+ | 0.122 | 4.81 | 90000 | 0.9483 | 0.8559 | 0.8319 |
90
 
91
 
92
  ### Framework versions
config.json CHANGED
@@ -1,12 +1,9 @@
1
  {
2
- "_name_or_path": "facebook/xlm-v-base",
3
  "architectures": [
4
- "XLMRobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
8
- "classifier_dropout": null,
9
- "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
@@ -136,17 +133,27 @@
136
  "LABEL_8": 8,
137
  "LABEL_9": 9
138
  },
139
- "layer_norm_eps": 1e-05,
140
- "max_position_embeddings": 514,
141
- "model_type": "xlm-roberta",
 
 
142
  "num_attention_heads": 12,
143
  "num_hidden_layers": 12,
144
- "pad_token_id": 1,
145
- "position_embedding_type": "absolute",
146
- "problem_type": "single_label_classification",
 
 
 
 
 
 
 
 
 
147
  "torch_dtype": "float32",
148
  "transformers_version": "4.33.3",
149
- "type_vocab_size": 1,
150
- "use_cache": true,
151
- "vocab_size": 901629
152
  }
 
1
  {
2
+ "_name_or_path": "microsoft/mdeberta-v3-base",
3
  "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
 
 
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 768,
 
133
  "LABEL_8": 8,
134
  "LABEL_9": 9
135
  },
136
+ "layer_norm_eps": 1e-07,
137
+ "max_position_embeddings": 512,
138
+ "max_relative_positions": -1,
139
+ "model_type": "deberta-v2",
140
+ "norm_rel_ebd": "layer_norm",
141
  "num_attention_heads": 12,
142
  "num_hidden_layers": 12,
143
+ "pad_token_id": 0,
144
+ "pooler_dropout": 0,
145
+ "pooler_hidden_act": "gelu",
146
+ "pooler_hidden_size": 768,
147
+ "pos_att_type": [
148
+ "p2c",
149
+ "c2p"
150
+ ],
151
+ "position_biased_input": false,
152
+ "position_buckets": 256,
153
+ "relative_attention": true,
154
+ "share_att_key": true,
155
  "torch_dtype": "float32",
156
  "transformers_version": "4.33.3",
157
+ "type_vocab_size": 0,
158
+ "vocab_size": 251000
 
159
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f71b68528e7dfc888fd501d031df288e55220a595fedd44e0a074c4489ea6c6
3
- size 3114226670
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46e4599edc4a5f0d569a174684dbef4435918fc6cb26d0986eb6eab29335e22a
3
+ size 1115491954
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfd5c5598e502bf60285c2298617fec9b1aa82429d2620bf1eaf88144f6746de
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5415336326de347ef5dd173e77a1892eb5faece5e0a9bdf316852ba765f25a5a
3
  size 4600