EYEDOL commited on
Commit
99cc2e5
1 Parent(s): 0343fe8

End of training

Browse files
README.md CHANGED
@@ -20,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 11.0 dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.1906
24
 
25
  ## Model description
26
 
@@ -48,15 +48,16 @@ The following hyperparameters were used during training:
48
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
  - lr_scheduler_type: linear
50
  - lr_scheduler_warmup_steps: 50
51
- - num_epochs: 2
52
  - mixed_precision_training: Native AMP
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss |
57
  |:-------------:|:-----:|:----:|:---------------:|
58
- | 0.2195 | 1.0 | 491 | 0.2474 |
59
- | 0.1199 | 2.0 | 982 | 0.1906 |
 
60
 
61
 
62
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 11.0 dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.1913
24
 
25
  ## Model description
26
 
 
48
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
  - lr_scheduler_type: linear
50
  - lr_scheduler_warmup_steps: 50
51
+ - num_epochs: 3
52
  - mixed_precision_training: Native AMP
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss |
57
  |:-------------:|:-----:|:----:|:---------------:|
58
+ | 0.2329 | 1.0 | 491 | 0.2707 |
59
+ | 0.1468 | 2.0 | 982 | 0.2118 |
60
+ | 0.0747 | 3.0 | 1473 | 0.1913 |
61
 
62
 
63
  ### Framework versions
adapter_config.json CHANGED
@@ -16,6 +16,7 @@
16
  "layers_to_transform": null,
17
  "loftq_config": {},
18
  "lora_alpha": 64,
 
19
  "lora_dropout": 0.05,
20
  "megatron_config": null,
21
  "megatron_core": "megatron.core",
@@ -25,8 +26,8 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "v_proj",
29
- "q_proj"
30
  ],
31
  "task_type": null,
32
  "use_dora": false,
 
16
  "layers_to_transform": null,
17
  "loftq_config": {},
18
  "lora_alpha": 64,
19
+ "lora_bias": false,
20
  "lora_dropout": 0.05,
21
  "megatron_config": null,
22
  "megatron_core": "megatron.core",
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "q_proj",
30
+ "v_proj"
31
  ],
32
  "task_type": null,
33
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83b4f827d895119dc8ff926dd60e1549f0bc431022d1c4db96aac06de0663971
3
  size 62969640
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c1a2e4c094cb776103dd5332b64d45fb5ad6e5fc6dce641329164399d953744
3
  size 62969640
runs/Nov27_20-01-59_e97c258ee6ab/events.out.tfevents.1732737723.e97c258ee6ab.23.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8b803e3fffcbb1a5ceaf877f3f24b600f0c1e1384df3fa2b84d99d3a27aa2df
3
+ size 19738
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c284c80e56bab7dd4e1e36f4106b13146be048abb62c98ea493629114e60106
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09dc24ab8d36c45c047bc52f3f022e4921d494599d30ab853221d8ca1a3d2772
3
  size 5496