anushaporwal commited on
Commit
e113e3a
1 Parent(s): b223f7f

Model save

Browse files
Files changed (3) hide show
  1. README.md +58 -0
  2. all_results.json +4 -4
  3. train_results.json +4 -4
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: facebook/wav2vec2-large-xlsr-53
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - common_voice_16_0
8
+ model-index:
9
+ - name: wav2vec2-common_voice-tr-demo-1
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # wav2vec2-common_voice-tr-demo-1
17
+
18
+ This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the common_voice_16_0 dataset.
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ ### Training hyperparameters
35
+
36
+ The following hyperparameters were used during training:
37
+ - learning_rate: 0.0003
38
+ - train_batch_size: 16
39
+ - eval_batch_size: 8
40
+ - seed: 42
41
+ - gradient_accumulation_steps: 2
42
+ - total_train_batch_size: 32
43
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
+ - lr_scheduler_type: linear
45
+ - lr_scheduler_warmup_steps: 500
46
+ - num_epochs: 3.0
47
+ - mixed_precision_training: Native AMP
48
+
49
+ ### Training results
50
+
51
+
52
+
53
+ ### Framework versions
54
+
55
+ - Transformers 4.43.0.dev0
56
+ - Pytorch 2.3.1+cu121
57
+ - Datasets 2.19.1
58
+ - Tokenizers 0.19.1
all_results.json CHANGED
@@ -6,8 +6,8 @@
6
  "eval_steps_per_second": 3.378,
7
  "eval_wer": 1.0,
8
  "total_flos": 3.8622091973981184e+17,
9
- "train_loss": 14.760249496788107,
10
- "train_runtime": 133.5145,
11
- "train_samples_per_second": 22.469,
12
- "train_steps_per_second": 0.697
13
  }
 
6
  "eval_steps_per_second": 3.378,
7
  "eval_wer": 1.0,
8
  "total_flos": 3.8622091973981184e+17,
9
+ "train_loss": 14.888988710218861,
10
+ "train_runtime": 132.1117,
11
+ "train_samples_per_second": 22.708,
12
+ "train_steps_per_second": 0.704
13
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 2.9523809523809526,
3
  "total_flos": 3.8622091973981184e+17,
4
- "train_loss": 14.760249496788107,
5
- "train_runtime": 133.5145,
6
- "train_samples_per_second": 22.469,
7
- "train_steps_per_second": 0.697
8
  }
 
1
  {
2
  "epoch": 2.9523809523809526,
3
  "total_flos": 3.8622091973981184e+17,
4
+ "train_loss": 14.888988710218861,
5
+ "train_runtime": 132.1117,
6
+ "train_samples_per_second": 22.708,
7
+ "train_steps_per_second": 0.704
8
  }