MehdiHosseiniMoghadam
commited on
Commit
•
7f8fcfe
1
Parent(s):
958dd26
Add model files
Browse files- config.json +76 -0
- preprocessor_config.json +8 -0
- pytorch_model.bin +3 -0
- trainer_state.json +0 -0
- training_args.bin +3 -0
- vocab.json +1 -0
config.json
ADDED
@@ -0,0 +1,76 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "facebook/wav2vec2-large-xlsr-53",
|
3 |
+
"activation_dropout": 0.0,
|
4 |
+
"apply_spec_augment": true,
|
5 |
+
"architectures": [
|
6 |
+
"Wav2Vec2ForCTC"
|
7 |
+
],
|
8 |
+
"attention_dropout": 0.1,
|
9 |
+
"bos_token_id": 1,
|
10 |
+
"conv_bias": true,
|
11 |
+
"conv_dim": [
|
12 |
+
512,
|
13 |
+
512,
|
14 |
+
512,
|
15 |
+
512,
|
16 |
+
512,
|
17 |
+
512,
|
18 |
+
512
|
19 |
+
],
|
20 |
+
"conv_kernel": [
|
21 |
+
10,
|
22 |
+
3,
|
23 |
+
3,
|
24 |
+
3,
|
25 |
+
3,
|
26 |
+
2,
|
27 |
+
2
|
28 |
+
],
|
29 |
+
"conv_stride": [
|
30 |
+
5,
|
31 |
+
2,
|
32 |
+
2,
|
33 |
+
2,
|
34 |
+
2,
|
35 |
+
2,
|
36 |
+
2
|
37 |
+
],
|
38 |
+
"ctc_loss_reduction": "mean",
|
39 |
+
"ctc_zero_infinity": false,
|
40 |
+
"do_stable_layer_norm": true,
|
41 |
+
"eos_token_id": 2,
|
42 |
+
"feat_extract_activation": "gelu",
|
43 |
+
"feat_extract_dropout": 0.0,
|
44 |
+
"feat_extract_norm": "layer",
|
45 |
+
"feat_proj_dropout": 0.0,
|
46 |
+
"final_dropout": 0.0,
|
47 |
+
"gradient_checkpointing": true,
|
48 |
+
"hidden_act": "gelu",
|
49 |
+
"hidden_dropout": 0.1,
|
50 |
+
"hidden_size": 1024,
|
51 |
+
"initializer_range": 0.02,
|
52 |
+
"intermediate_size": 4096,
|
53 |
+
"layer_norm_eps": 1e-05,
|
54 |
+
"layerdrop": 0.1,
|
55 |
+
"mask_channel_length": 10,
|
56 |
+
"mask_channel_min_space": 1,
|
57 |
+
"mask_channel_other": 0.0,
|
58 |
+
"mask_channel_prob": 0.0,
|
59 |
+
"mask_channel_selection": "static",
|
60 |
+
"mask_feature_length": 10,
|
61 |
+
"mask_feature_prob": 0.0,
|
62 |
+
"mask_time_length": 10,
|
63 |
+
"mask_time_min_space": 1,
|
64 |
+
"mask_time_other": 0.0,
|
65 |
+
"mask_time_prob": 0.05,
|
66 |
+
"mask_time_selection": "static",
|
67 |
+
"model_type": "wav2vec2",
|
68 |
+
"num_attention_heads": 16,
|
69 |
+
"num_conv_pos_embedding_groups": 16,
|
70 |
+
"num_conv_pos_embeddings": 128,
|
71 |
+
"num_feat_extract_layers": 7,
|
72 |
+
"num_hidden_layers": 24,
|
73 |
+
"pad_token_id": 185,
|
74 |
+
"transformers_version": "4.4.0",
|
75 |
+
"vocab_size": 186
|
76 |
+
}
|
preprocessor_config.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"do_normalize": true,
|
3 |
+
"feature_size": 1,
|
4 |
+
"padding_side": "right",
|
5 |
+
"padding_value": 0.0,
|
6 |
+
"return_attention_mask": true,
|
7 |
+
"sampling_rate": 16000
|
8 |
+
}
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9f289ae48c8a9772e811ee23ed5c4b2244138abc3e731a99274343a5794db11c
|
3 |
+
size 1262692022
|
trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:43f1a946892bde03753c904a631cc0aeacb0e52a89776a81ee103adf3c23bcad
|
3 |
+
size 2287
|
vocab.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"\u010f": 0, "x": 1, "h": 2, "\u81e3": 3, "\u00a1": 4, "\u0113": 5, "t": 6, "\u043e": 7, "\u00e0": 8, "\u00e9": 9, "\u5b59": 10, "\u01d4": 11, "\u0447": 12, "\u00ed": 13, "\u0306": 14, "]": 15, "\u0430": 16, "[": 17, "\u00ef": 18, "\u0259": 19, "\u0432": 20, "o": 21, "\u0119": 22, "\u2013": 23, "\u0121": 24, "\u00bb": 25, "\u0444": 26, "\u015f": 27, "\u0448": 28, "$": 29, "\u4e61": 30, "\u0161": 31, "\u0101": 32, "\u5e7a": 33, "_": 34, "(": 35, "\u015b": 36, "\u01d0": 37, "\u0441": 38, "\u0440": 39, "\u0146": 40, "\u1ed9": 41, "\u30ab": 42, "\u00f3": 43, "\u017e": 44, "\u00fb": 45, "\u203a": 46, "\u2011": 47, "\u1ea1": 48, "\u02bf": 50, "\u00e7": 51, "i": 52, "'": 53, "\u1ec5": 54, "\u021b": 55, "\u00fc": 56, "\u0435": 57, "c": 58, "\u0117": 59, "\u02bb": 60, "n": 61, "\u0f0b": 62, "u": 63, "\u043a": 64, "\u2033": 65, "\u00b5": 66, "\u2014": 67, "\u1ed3": 68, "\u011b": 69, "\u00ea": 70, "f": 71, "\u0103": 72, "\u0165": 73, "\u0105": 74, "\u014d": 75, "\u0142": 76, "q": 77, "\u00f2": 78, "\u05e2": 79, "\u00b4": 80, "&": 81, "\u6bd4": 82, "\u0148": 83, "\u00e1": 84, "\u012b": 85, "e": 86, "\u00ab": 87, "\u0144": 88, "\u00e5": 89, "\u0438": 90, "\u201f": 91, "\u1ea3": 92, "j": 93, "\u0131": 94, "\u0479": 95, ")": 96, "\u1ebf": 97, "+": 98, "/": 99, "\u00fe": 100, "\u00f6": 101, "\u0153": 102, "b": 103, "\u2032": 104, "\u1ee9": 105, "\u2212": 106, "v": 107, "\u7121": 108, "\u00ee": 109, "\u1e5f": 110, "\u00b0": 111, "\u00ec": 112, "\u201e": 113, "\u00f0": 114, "\u00fd": 115, "\u05d1": 116, "a": 117, "g": 118, "\u00fa": 119, "\u4e34": 120, "\u1e2b": 121, "\u00e6": 122, "\u05e9": 123, "\u1eb1": 124, "\u0151": 125, "\u0127": 126, "r": 127, "\u016b": 128, "w": 129, "\u652f": 130, "y": 131, "\u0159": 132, "\u0f53": 133, "\u2026": 134, "\u00f1": 135, "\u751f": 136, "k": 137, "\u00f9": 138, "\u011f": 139, "s": 140, "\u05e0": 141, "\u0111": 142, "\u8fb6": 143, "\u016f": 144, "\u00e2": 145, "\u00f8": 146, "@": 147, "\u014f": 148, "\u5c23": 149, "\u2192": 150, "\u1eaf": 151, "\u27e9": 152, "\u0307": 153, "\u0219": 154, "\u00df": 155, "\u00b7": 156, "`": 157, "\u017c": 158, "d": 159, "\u00e4": 160, "\u201a": 161, "z": 162, "m": 163, "\u1e6d": 164, "\u1ecd": 165, "\u05d0": 166, "\u017a": 167, "\u010d": 168, "\u2039": 169, "\u0107": 170, "\u00f5": 171, "\u00e3": 172, "=": 173, "\u27e8": 174, "\u00f4": 175, "\u1e63": 176, "\u2261": 177, "\u2019": 178, "l": 179, "\u043c": 180, "\u00eb": 181, "\u9053": 182, "p": 183, "|": 49, "[UNK]": 184, "[PAD]": 185}
|