quinnb commited on
Commit
08b88ef
1 Parent(s): 9c74ba4

End of training

Browse files
Files changed (4) hide show
  1. README.md +5 -12
  2. config.json +1 -3
  3. generation_config.json +2 -10
  4. training_args.bin +2 -2
README.md CHANGED
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # Whisper Large v3 Trained on Hindi
19
 
20
- This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the Custom Hindi dataset dataset.
21
 
22
  ## Model description
23
 
@@ -37,24 +37,17 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 1e-05
40
- - train_batch_size: 1
41
- - eval_batch_size: 16
42
  - seed: 42
43
- - distributed_type: multi-GPU
44
- - num_devices: 4
45
  - gradient_accumulation_steps: 16
46
  - total_train_batch_size: 64
47
- - total_eval_batch_size: 64
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
- - lr_scheduler_warmup_steps: 250
51
- - num_epochs: 3.0
52
  - mixed_precision_training: Native AMP
53
 
54
- ### Training results
55
-
56
-
57
-
58
  ### Framework versions
59
 
60
  - Transformers 4.41.1
 
17
 
18
  # Whisper Large v3 Trained on Hindi
19
 
20
+ This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the Common Voice 17.0 dataset.
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 1e-05
40
+ - train_batch_size: 4
41
+ - eval_batch_size: 64
42
  - seed: 42
 
 
43
  - gradient_accumulation_steps: 16
44
  - total_train_batch_size: 64
 
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
+ - lr_scheduler_warmup_steps: 500
48
+ - training_steps: 20
49
  - mixed_precision_training: Native AMP
50
 
 
 
 
 
51
  ### Framework versions
52
 
53
  - Transformers 4.41.1
config.json CHANGED
@@ -25,7 +25,6 @@
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 32,
27
  "eos_token_id": 50257,
28
- "forced_decoder_ids": null,
29
  "init_std": 0.02,
30
  "is_encoder_decoder": true,
31
  "mask_feature_length": 10,
@@ -43,8 +42,7 @@
43
  "num_mel_bins": 128,
44
  "pad_token_id": 50256,
45
  "scale_embedding": false,
46
- "suppress_tokens": [],
47
- "torch_dtype": "float16",
48
  "transformers_version": "4.41.1",
49
  "use_cache": true,
50
  "use_weighted_layer_sum": false,
 
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 32,
27
  "eos_token_id": 50257,
 
28
  "init_std": 0.02,
29
  "is_encoder_decoder": true,
30
  "mask_feature_length": 10,
 
42
  "num_mel_bins": 128,
43
  "pad_token_id": 50256,
44
  "scale_embedding": false,
45
+ "torch_dtype": "float32",
 
46
  "transformers_version": "4.41.1",
47
  "use_cache": true,
48
  "use_weighted_layer_sum": false,
generation_config.json CHANGED
@@ -48,16 +48,6 @@
48
  "bos_token_id": 50257,
49
  "decoder_start_token_id": 50258,
50
  "eos_token_id": 50257,
51
- "forced_decoder_ids": [
52
- [
53
- 1,
54
- null
55
- ],
56
- [
57
- 2,
58
- 50360
59
- ]
60
- ],
61
  "is_multilingual": true,
62
  "lang_to_id": {
63
  "<|af|>": 50327,
@@ -161,6 +151,7 @@
161
  "<|yue|>": 50358,
162
  "<|zh|>": 50260
163
  },
 
164
  "max_initial_timestamp_index": 50,
165
  "max_length": 448,
166
  "no_timestamps_token_id": 50364,
@@ -257,6 +248,7 @@
257
  50362,
258
  50363
259
  ],
 
260
  "task_to_id": {
261
  "transcribe": 50360,
262
  "translate": 50359
 
48
  "bos_token_id": 50257,
49
  "decoder_start_token_id": 50258,
50
  "eos_token_id": 50257,
 
 
 
 
 
 
 
 
 
 
51
  "is_multilingual": true,
52
  "lang_to_id": {
53
  "<|af|>": 50327,
 
151
  "<|yue|>": 50358,
152
  "<|zh|>": 50260
153
  },
154
+ "language": "hindi",
155
  "max_initial_timestamp_index": 50,
156
  "max_length": 448,
157
  "no_timestamps_token_id": 50364,
 
248
  50362,
249
  50363
250
  ],
251
+ "task": "transcribe",
252
  "task_to_id": {
253
  "transcribe": 50360,
254
  "translate": 50359
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3808877e186e53d02403c6cdca7d73cd3a22cc5292819e042fa3faa5156d51e0
3
- size 5615
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34facea37d4705d15a4fdc1db99d081ec602439fc3e74151978af6c6ac17b08e
3
+ size 4783