ad019el commited on
Commit
c7fc1d7
1 Parent(s): 630a7a0

End Training

Browse files
README.md ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: ad019el/tamasheq-99-final
3
+ datasets:
4
+ - ad019el/ar_data
5
+ - heisenberg1337/tamasheq_data
6
+ metrics:
7
+ - cer
8
+ - wer
9
+ tags:
10
+ - generated_from_trainer
11
+ ---
12
+ model-index:
13
+ - name: tamasheq-99-final
14
+ results: []
15
+ ---
16
+
17
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
+ should probably proofread and complete it, then remove this comment. -->
19
+
20
+ # tamasheq-99-final
21
+
22
+ This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-arabic](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-arabic) on the None dataset.
23
+ It achieves the following results on the evaluation set:
24
+ - Cer: 16.2959
25
+ - Wer: 55.5334
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 3e-05
45
+ - train_batch_size: 16
46
+ - eval_batch_size: 8
47
+ - seed: 42
48
+ - gradient_accumulation_steps: 2
49
+ - total_train_batch_size: 32
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: linear
52
+ - lr_scheduler_warmup_steps: 500
53
+
54
+ ### Training results
55
+
56
+ |step |tamasheq_wer|arabic_wer|tamasheq_cer|arabic_cer|
57
+ |------------|------------|----------|------------|----------|
58
+ |Before train|104.985 |23.1305 |67.4458 |7.30972 |
59
+ |step 300 |99.5513 |23.0544 |49.7078 |7.1043 |
60
+ |step 600 |95.1147 |22.5267 |41.4515 |6.0098 |
61
+ |step 900 |93.5194 |21.0404 |38.0867 |5.52939 |
62
+ |step 1200 |92.5723 |20.6224 |37.0877 |5.39751 |
63
+ |step 1500 |92.3009 |20.9238 |36.9915 |5.6718 |
64
+ |step 1800 |92.0738 |21.2699 |36.3713 |6.08877 |
65
+ |step 2100 |88.7338 |21.9693 |33.3648 |5.9156 |
66
+ |step 2400 |87.1884 |21.1333 |31.8379 |5.52939 |
67
+ |step 2700 |88.299 |21.0705 |31.4599 |5.5078 |
68
+ |step 3000 |87.7866 |21.5021 |30.9039 |6.29239 |
69
+ |step 3300 |84.2971 |21.666 |29.7455 |5.97212 |
70
+ |step 3600 |83.8983 |21.5732 |28.6145 |6.04748 |
71
+ |step 3900 |81.8544 |22.1087 |27.9359 |5.99096 |
72
+ |step 4200 |82.9741 |23.392 |27.4288 |6.4013 |
73
+ |step 4500 |83.8485 |24.2452 |27.0575 |6.79164 |
74
+ |step 4800 |81.6052 |22.666 |26.6918 |6.09457 |
75
+ |step 5100 |77.9661 |22.4803 |25.1084 |6.0098 |
76
+ |step 5400 |77.2183 |21.83 |24.656 |5.9156 |
77
+ |step 5700 |76.672 |22.1078 |24.2606 |6.0802 |
78
+ |step 6000 |76.2712 |22.7589 |23.9236 |6.41485 |
79
+ |step 6300 |75.7228 |23.8737 |23.7135 |6.78222 |
80
+ |step 6600 |71.2363 |23.177 |22.196 |6.39601 |
81
+ |step 6900 |69.8405 |22.7125 |21.574 |6.21703 |
82
+ |step 7200 |72.9452 |23.6679 |21.0775 |6.6918 |
83
+ |step 7500 |75.9222 |24.7097 |20.8999 |7.17784 |
84
+ |step 7800 |67.4975 |23.1305 |20.6786 |6.65034 |
85
+ |step 8100 |65.2542 |23.1305 |19.7361 |6.49962 |
86
+ |step 8400 |61.7149 |22.3874 |18.426 |6.12283 |
87
+ |step 8700 |63.8046 |23.6679 |18.2166 |6.2679 |
88
+ |step 9000 |64.7059 |24.1059 |17.9952 |6.66918 |
89
+ |step 9300 |67.5474 |24.7097 |17.6078 |7.16843 |
90
+ |step 9600 |57.1286 |23.3163 |17.2385 |6.66918 |
91
+ |step 9900 |58.2752 |22.8054 |17.1065 |6.4431 |
92
+ |step 10200 |57.7767 |24.2917 |16.848 |6.68802 |
93
+ |step 10500 |55.2841 |25.1277 |16.5033 |7.12133 |
94
+ |step 10800 |52.5424 |23.8272 |15.9566 |6.80106 |
95
+ |step 11100 |55.5334 |24.6168 |16.2959 |6.94235 |
96
+
97
+ ### Framework versions
98
+
99
+ - Transformers 4.31.0
100
+ - Pytorch 2.0.1+cu118
101
+ - Datasets 2.14.4
102
+ - Tokenizers 0.13.3
config.json ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "ad019el/tamasheq-99-6",
3
+ "activation_dropout": 0.05,
4
+ "adapter_attn_dim": null,
5
+ "adapter_kernel_size": 3,
6
+ "adapter_stride": 2,
7
+ "add_adapter": false,
8
+ "apply_spec_augment": true,
9
+ "architectures": [
10
+ "Wav2Vec2ForCTC"
11
+ ],
12
+ "attention_dropout": 0.1,
13
+ "bos_token_id": 1,
14
+ "classifier_proj_size": 256,
15
+ "codevector_dim": 256,
16
+ "contrastive_logits_temperature": 0.1,
17
+ "conv_bias": true,
18
+ "conv_dim": [
19
+ 512,
20
+ 512,
21
+ 512,
22
+ 512,
23
+ 512,
24
+ 512,
25
+ 512
26
+ ],
27
+ "conv_kernel": [
28
+ 10,
29
+ 3,
30
+ 3,
31
+ 3,
32
+ 3,
33
+ 2,
34
+ 2
35
+ ],
36
+ "conv_stride": [
37
+ 5,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 2,
42
+ 2,
43
+ 2
44
+ ],
45
+ "ctc_loss_reduction": "mean",
46
+ "ctc_zero_infinity": true,
47
+ "diversity_loss_weight": 0.1,
48
+ "do_stable_layer_norm": true,
49
+ "eos_token_id": 2,
50
+ "feat_extract_activation": "gelu",
51
+ "feat_extract_dropout": 0.0,
52
+ "feat_extract_norm": "layer",
53
+ "feat_proj_dropout": 0.0,
54
+ "feat_quantizer_dropout": 0.0,
55
+ "final_dropout": 0.0,
56
+ "hidden_act": "gelu",
57
+ "hidden_dropout": 0.1,
58
+ "hidden_size": 1024,
59
+ "initializer_range": 0.02,
60
+ "intermediate_size": 4096,
61
+ "layer_norm_eps": 1e-05,
62
+ "layerdrop": 0.1,
63
+ "mask_channel_length": 10,
64
+ "mask_channel_min_space": 1,
65
+ "mask_channel_other": 0.0,
66
+ "mask_channel_prob": 0.0,
67
+ "mask_channel_selection": "static",
68
+ "mask_feature_length": 10,
69
+ "mask_feature_min_masks": 0,
70
+ "mask_feature_prob": 0.0,
71
+ "mask_time_length": 10,
72
+ "mask_time_min_masks": 2,
73
+ "mask_time_min_space": 1,
74
+ "mask_time_other": 0.0,
75
+ "mask_time_prob": 0.05,
76
+ "mask_time_selection": "static",
77
+ "model_type": "wav2vec2",
78
+ "num_adapter_layers": 3,
79
+ "num_attention_heads": 16,
80
+ "num_codevector_groups": 2,
81
+ "num_codevectors_per_group": 320,
82
+ "num_conv_pos_embedding_groups": 16,
83
+ "num_conv_pos_embeddings": 128,
84
+ "num_feat_extract_layers": 7,
85
+ "num_hidden_layers": 24,
86
+ "num_negatives": 100,
87
+ "output_hidden_size": 1024,
88
+ "pad_token_id": 42,
89
+ "proj_codevector_dim": 256,
90
+ "tdnn_dilation": [
91
+ 1,
92
+ 2,
93
+ 3,
94
+ 1,
95
+ 1
96
+ ],
97
+ "tdnn_dim": [
98
+ 512,
99
+ 512,
100
+ 512,
101
+ 512,
102
+ 1500
103
+ ],
104
+ "tdnn_kernel": [
105
+ 5,
106
+ 3,
107
+ 3,
108
+ 1,
109
+ 1
110
+ ],
111
+ "torch_dtype": "float32",
112
+ "transformers_version": "4.31.0",
113
+ "use_weighted_layer_sum": false,
114
+ "vocab_size": 43,
115
+ "xvector_output_dim": 512
116
+ }
preprocessor_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
+ "feature_size": 1,
5
+ "padding_side": "right",
6
+ "padding_value": 0.0,
7
+ "return_attention_mask": true,
8
+ "sampling_rate": 16000
9
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba0dd02b2e46c1cb5a1d244f1d1a10445e9a367f62825ffca69610376406b485
3
+ size 1262078125
runs/Aug15_17-53-51_f163a2af2bd6/events.out.tfevents.1692122092.f163a2af2bd6.3941.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f7dec8c4a9d958b2b98823f254239e811b1bbd0eccb30f406f52757a7db812c
3
+ size 7580
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "[PAD]",
5
+ "unk_token": "[UNK]"
6
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "clean_up_tokenization_spaces": true,
4
+ "do_lower_case": false,
5
+ "eos_token": "</s>",
6
+ "model_max_length": 1000000000000000019884624838656,
7
+ "pad_token": "[PAD]",
8
+ "replace_word_delimiter_char": " ",
9
+ "target_lang": null,
10
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
11
+ "unk_token": "[UNK]",
12
+ "word_delimiter_token": "|"
13
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb96442376da9c29e07c372565529befc21db5ba6630822b1a3b2ced5f51ccfb
3
+ size 3963
vocab.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "[PAD]": 42,
3
+ "[UNK]": 41,
4
+ "|": 22,
5
+ "ء": 38,
6
+ "آ": 6,
7
+ "أ": 10,
8
+ "ؤ": 13,
9
+ "إ": 27,
10
+ "ئ": 12,
11
+ "ا": 14,
12
+ "ب": 18,
13
+ "ة": 5,
14
+ "ت": 26,
15
+ "ث": 25,
16
+ "ج": 35,
17
+ "ح": 17,
18
+ "خ": 28,
19
+ "د": 8,
20
+ "ذ": 21,
21
+ "ر": 16,
22
+ "ز": 29,
23
+ "س": 39,
24
+ "ش": 19,
25
+ "ص": 36,
26
+ "ض": 20,
27
+ "ط": 1,
28
+ "ظ": 37,
29
+ "ع": 0,
30
+ "غ": 33,
31
+ "ف": 7,
32
+ "ق": 15,
33
+ "ك": 24,
34
+ "ل": 11,
35
+ "م": 31,
36
+ "ن": 30,
37
+ "ه": 2,
38
+ "و": 4,
39
+ "ى": 23,
40
+ "ي": 3,
41
+ "ٱ": 32,
42
+ "ڤ": 9,
43
+ "ک": 40,
44
+ "ی": 34
45
+ }