Upload WhisperForConditionalGeneration

#4
Files changed (4) hide show
  1. README.md +2 -2
  2. config.json +3 -2
  3. generation_config.json +6 -9
  4. model.safetensors +3 -0
README.md CHANGED
@@ -2,10 +2,10 @@
2
  language:
3
  - hi
4
  license: apache-2.0
5
- tags:
6
- - whisper-event
7
  metrics:
8
  - wer
 
 
9
  model-index:
10
  - name: Whisper Hindi Medium - Vasista Sai Lodagala
11
  results:
 
2
  language:
3
  - hi
4
  license: apache-2.0
 
 
5
  metrics:
6
  - wer
7
+ tags:
8
+ - whisper-event
9
  model-index:
10
  - name: Whisper Hindi Medium - Vasista Sai Lodagala
11
  results:
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "openai/whisper-medium",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
@@ -37,6 +37,7 @@
37
  "max_length": 448,
38
  "max_source_positions": 1500,
39
  "max_target_positions": 448,
 
40
  "model_type": "whisper",
41
  "num_hidden_layers": 24,
42
  "num_mel_bins": 80,
@@ -44,7 +45,7 @@
44
  "scale_embedding": false,
45
  "suppress_tokens": [],
46
  "torch_dtype": "float32",
47
- "transformers_version": "4.27.4",
48
  "use_cache": false,
49
  "use_weighted_layer_sum": false,
50
  "vocab_size": 51865
 
1
  {
2
+ "_name_or_path": "vasista22/whisper-hindi-medium",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
37
  "max_length": 448,
38
  "max_source_positions": 1500,
39
  "max_target_positions": 448,
40
+ "median_filter_width": 7,
41
  "model_type": "whisper",
42
  "num_hidden_layers": 24,
43
  "num_mel_bins": 80,
 
45
  "scale_embedding": false,
46
  "suppress_tokens": [],
47
  "torch_dtype": "float32",
48
+ "transformers_version": "4.42.4",
49
  "use_cache": false,
50
  "use_weighted_layer_sum": false,
51
  "vocab_size": 51865
generation_config.json CHANGED
@@ -1,15 +1,12 @@
1
  {
2
- "_from_model_config": true,
3
- "begin_suppress_tokens": [
4
- 220,
5
- 50257
6
- ],
7
  "bos_token_id": 50257,
8
- "decoder_start_token_id": 50258,
9
  "eos_token_id": 50257,
 
10
  "max_length": 448,
 
 
 
11
  "pad_token_id": 50257,
12
- "suppress_tokens": [],
13
- "transformers_version": "4.27.4",
14
- "use_cache": false
15
  }
 
1
  {
 
 
 
 
 
2
  "bos_token_id": 50257,
3
+ "early_stopping": true,
4
  "eos_token_id": 50257,
5
+ "length_penalty": 2.0,
6
  "max_length": 448,
7
+ "no_repeat_ngram_size": 3,
8
+ "no_timestamps_token_id": 50256,
9
+ "num_beams": 4,
10
  "pad_token_id": 50257,
11
+ "transformers_version": "4.42.4"
 
 
12
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:317ec66e3ddfcc7e5c61b2f1f8d4e6338b146c62910c14a54267b64049c794ea
3
+ size 3055544304