hongjing0312 commited on
Commit
50a03e4
1 Parent(s): 4fddfa3

Training in progress, step 500

Browse files
README.md CHANGED
@@ -3,26 +3,11 @@ license: apache-2.0
3
  base_model: t5-small
4
  tags:
5
  - generated_from_trainer
6
- datasets:
7
- - opus_books
8
  metrics:
9
  - bleu
10
  model-index:
11
  - name: my_awesome_opus_books_model
12
- results:
13
- - task:
14
- name: Sequence-to-sequence Language Modeling
15
- type: text2text-generation
16
- dataset:
17
- name: opus_books
18
- type: opus_books
19
- config: en-es
20
- split: train
21
- args: en-es
22
- metrics:
23
- - name: Bleu
24
- type: bleu
25
- value: 0.8718
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -30,11 +15,11 @@ should probably proofread and complete it, then remove this comment. -->
30
 
31
  # my_awesome_opus_books_model
32
 
33
- This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the opus_books dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 2.1549
36
- - Bleu: 0.8718
37
- - Gen Len: 18.1062
38
 
39
  ## Model description
40
 
@@ -59,26 +44,23 @@ The following hyperparameters were used during training:
59
  - seed: 42
60
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
  - lr_scheduler_type: linear
62
- - num_epochs: 8
63
  - mixed_precision_training: Native AMP
64
 
65
  ### Training results
66
 
67
  | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
68
  |:-------------:|:-----:|:-----:|:---------------:|:------:|:-------:|
69
- | 2.9712 | 1.0 | 4674 | 2.6344 | 0.3403 | 18.2053 |
70
- | 2.7554 | 2.0 | 9348 | 2.4373 | 0.4973 | 18.1724 |
71
- | 2.6295 | 3.0 | 14022 | 2.3271 | 0.638 | 18.1463 |
72
- | 2.5452 | 4.0 | 18696 | 2.2553 | 0.7342 | 18.1319 |
73
- | 2.4968 | 5.0 | 23370 | 2.2089 | 0.7964 | 18.1238 |
74
- | 2.4531 | 6.0 | 28044 | 2.1771 | 0.8458 | 18.1048 |
75
- | 2.4441 | 7.0 | 32718 | 2.1605 | 0.8637 | 18.1065 |
76
- | 2.42 | 8.0 | 37392 | 2.1549 | 0.8718 | 18.1062 |
77
 
78
 
79
  ### Framework versions
80
 
81
- - Transformers 4.36.2
82
- - Pytorch 2.1.2
83
  - Datasets 2.15.0
84
  - Tokenizers 0.15.0
 
3
  base_model: t5-small
4
  tags:
5
  - generated_from_trainer
 
 
6
  metrics:
7
  - bleu
8
  model-index:
9
  - name: my_awesome_opus_books_model
10
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
15
 
16
  # my_awesome_opus_books_model
17
 
18
+ This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 2.2384
21
+ - Bleu: 0.4736
22
+ - Gen Len: 18.0826
23
 
24
  ## Model description
25
 
 
44
  - seed: 42
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
+ - num_epochs: 5
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
53
  |:-------------:|:-----:|:-----:|:---------------:|:------:|:-------:|
54
+ | 2.9148 | 1.0 | 4674 | 2.5719 | 0.2155 | 18.2619 |
55
+ | 2.7067 | 2.0 | 9348 | 2.3872 | 0.3419 | 18.1173 |
56
+ | 2.5969 | 3.0 | 14022 | 2.2973 | 0.4065 | 18.101 |
57
+ | 2.5413 | 4.0 | 18696 | 2.2518 | 0.4603 | 18.0815 |
58
+ | 2.5223 | 5.0 | 23370 | 2.2384 | 0.4736 | 18.0826 |
 
 
 
59
 
60
 
61
  ### Framework versions
62
 
63
+ - Transformers 4.35.2
64
+ - Pytorch 2.1.1
65
  - Datasets 2.15.0
66
  - Tokenizers 0.15.0
config.json CHANGED
@@ -55,7 +55,7 @@
55
  }
56
  },
57
  "torch_dtype": "float32",
58
- "transformers_version": "4.36.2",
59
  "use_cache": true,
60
  "vocab_size": 32128
61
  }
 
55
  }
56
  },
57
  "torch_dtype": "float32",
58
+ "transformers_version": "4.35.2",
59
  "use_cache": true,
60
  "vocab_size": 32128
61
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
- "transformers_version": "4.36.2"
6
  }
 
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
+ "transformers_version": "4.35.2"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fd3620aea053651e236661a9f9c1bdc02f4c394f3eb5ba7130c2102225680c0
3
  size 242041896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7eb546d53b4cd5a535bc16e1dda74f157bd93aabed8ecf85f7d1cff5d2c7b738
3
  size 242041896
runs/Dec21_08-03-30_WJHF67-129/events.out.tfevents.1703117012.WJHF67-129.3916.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:987cd951e5331032b8bd279acb7442f07d64f2e1a3ff540771d34229202208b4
3
+ size 5380
special_tokens_map.json CHANGED
@@ -101,25 +101,7 @@
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
- "eos_token": {
105
- "content": "</s>",
106
- "lstrip": false,
107
- "normalized": false,
108
- "rstrip": false,
109
- "single_word": false
110
- },
111
- "pad_token": {
112
- "content": "<pad>",
113
- "lstrip": false,
114
- "normalized": false,
115
- "rstrip": false,
116
- "single_word": false
117
- },
118
- "unk_token": {
119
- "content": "<unk>",
120
- "lstrip": false,
121
- "normalized": false,
122
- "rstrip": false,
123
- "single_word": false
124
- }
125
  }
 
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
+ "eos_token": "</s>",
105
+ "pad_token": "<pad>",
106
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
107
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4186122426cd9077379a3ee8074cf02939f121a2230cd91887ce80ec7bfbfd1d
3
- size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b28647bbad4adf2887f44fd5a88c9c012cc315473350a058131fe0e1bbc7c06
3
+ size 4728