README.md CHANGED
@@ -1,28 +1,58 @@
1
  ---
2
  annotations_creators:
3
  - expert-generated
4
- language:
5
- - es
6
  language_creators:
7
  - other
 
 
8
  license:
9
  - cc-by-sa-4.0
10
  multilinguality:
11
  - monolingual
12
- pretty_name: 'CIEMPIESS BALANCE CORPUS: Audio and Transcripts of Mexican Spanish Broadcast Conversations.'
13
  size_categories:
14
  - 10K<n<100K
15
  source_datasets:
16
  - original
 
 
 
 
 
17
  tags:
18
  - ciempiess
19
  - spanish
20
  - mexican spanish
21
  - ciempiess project
22
  - ciempiess-unam project
23
- task_categories:
24
- - automatic-speech-recognition
25
- task_ids: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  ---
27
 
28
  # Dataset Card for ciempiess_balance
 
1
  ---
2
  annotations_creators:
3
  - expert-generated
 
 
4
  language_creators:
5
  - other
6
+ language:
7
+ - es
8
  license:
9
  - cc-by-sa-4.0
10
  multilinguality:
11
  - monolingual
 
12
  size_categories:
13
  - 10K<n<100K
14
  source_datasets:
15
  - original
16
+ task_categories:
17
+ - automatic-speech-recognition
18
+ task_ids: []
19
+ pretty_name: 'CIEMPIESS BALANCE CORPUS: Audio and Transcripts of Mexican Spanish Broadcast
20
+ Conversations.'
21
  tags:
22
  - ciempiess
23
  - spanish
24
  - mexican spanish
25
  - ciempiess project
26
  - ciempiess-unam project
27
+ dataset_info:
28
+ config_name: ciempiess_balance
29
+ features:
30
+ - name: audio_id
31
+ dtype: string
32
+ - name: audio
33
+ dtype:
34
+ audio:
35
+ sampling_rate: 16000
36
+ - name: speaker_id
37
+ dtype: string
38
+ - name: gender
39
+ dtype: string
40
+ - name: duration
41
+ dtype: float32
42
+ - name: normalized_text
43
+ dtype: string
44
+ splits:
45
+ - name: train
46
+ num_bytes: 1125593189.655
47
+ num_examples: 8555
48
+ download_size: 1295436525
49
+ dataset_size: 1125593189.655
50
+ configs:
51
+ - config_name: ciempiess_balance
52
+ data_files:
53
+ - split: train
54
+ path: ciempiess_balance/train-*
55
+ default: true
56
  ---
57
 
58
  # Dataset Card for ciempiess_balance
ciempiess_balance.py DELETED
@@ -1,123 +0,0 @@
1
- from collections import defaultdict
2
- import os
3
- import json
4
- import csv
5
-
6
- import datasets
7
-
8
- _NAME="ciempiess_balance"
9
- _VERSION="1.0.0"
10
- _AUDIO_EXTENSIONS=".flac"
11
-
12
- _DESCRIPTION = """
13
- CIEMPIESS BALANCE is a Radio Corpus designed to create acoustic models for automatic speech recognition. It is "balance" because it is designed to balance the CIEMPIESS LIGHT, which means that if both corpora are combined, one will get a gender balanced corpus.
14
- """
15
-
16
- _CITATION = """
17
- @misc{carlosmenaciempiessbalance2018,
18
- title={CIEMPIESS BALANCE CORPUS: Audio and Transcripts of Mexican Spanish Broadcast Conversations.},
19
- ldc_catalog_no={LDC2018S11},
20
- DOI={https://doi.org/10.35111/rfmw-n126},
21
- author={Hernandez Mena, Carlos Daniel},
22
- journal={Linguistic Data Consortium, Philadelphia},
23
- year={2018},
24
- url={https://catalog.ldc.upenn.edu/LDC2018S11},
25
- }
26
- """
27
-
28
- _HOMEPAGE = "https://catalog.ldc.upenn.edu/LDC2018S11"
29
-
30
- _LICENSE = "CC-BY-SA-4.0, See https://creativecommons.org/licenses/by-sa/4.0/"
31
-
32
- _BASE_DATA_DIR = "corpus/"
33
- _METADATA_TRAIN = os.path.join(_BASE_DATA_DIR,"files", "metadata_train.tsv")
34
-
35
- _TARS_TRAIN = os.path.join(_BASE_DATA_DIR,"files", "tars_train.paths")
36
-
37
- class CiempiessBalanceConfig(datasets.BuilderConfig):
38
- """BuilderConfig for CIEMPIESS BALANCE Corpus"""
39
-
40
- def __init__(self, name, **kwargs):
41
- name=_NAME
42
- super().__init__(name=name, **kwargs)
43
-
44
- class CiempiessBalance(datasets.GeneratorBasedBuilder):
45
- """CIEMPIESS BALANCE Corpus"""
46
-
47
- VERSION = datasets.Version(_VERSION)
48
- BUILDER_CONFIGS = [
49
- CiempiessBalanceConfig(
50
- name=_NAME,
51
- version=datasets.Version(_VERSION),
52
- )
53
- ]
54
-
55
- def _info(self):
56
- features = datasets.Features(
57
- {
58
- "audio_id": datasets.Value("string"),
59
- "audio": datasets.Audio(sampling_rate=16000),
60
- "speaker_id": datasets.Value("string"),
61
- "gender": datasets.Value("string"),
62
- "duration": datasets.Value("float32"),
63
- "normalized_text": datasets.Value("string"),
64
- }
65
- )
66
- return datasets.DatasetInfo(
67
- description=_DESCRIPTION,
68
- features=features,
69
- homepage=_HOMEPAGE,
70
- license=_LICENSE,
71
- citation=_CITATION,
72
- )
73
-
74
- def _split_generators(self, dl_manager):
75
-
76
- metadata_train=dl_manager.download_and_extract(_METADATA_TRAIN)
77
-
78
- tars_train=dl_manager.download_and_extract(_TARS_TRAIN)
79
-
80
- hash_tar_files=defaultdict(dict)
81
-
82
- with open(tars_train,'r') as f:
83
- hash_tar_files['train']=[path.replace('\n','') for path in f]
84
-
85
- hash_meta_paths={"train":metadata_train}
86
- audio_paths = dl_manager.download(hash_tar_files)
87
-
88
- splits=["train"]
89
- local_extracted_audio_paths = (
90
- dl_manager.extract(audio_paths) if not dl_manager.is_streaming else
91
- {
92
- split:[None] * len(audio_paths[split]) for split in splits
93
- }
94
- )
95
-
96
- return [
97
- datasets.SplitGenerator(
98
- name=datasets.Split.TRAIN,
99
- gen_kwargs={
100
- "audio_archives": [dl_manager.iter_archive(archive) for archive in audio_paths["train"]],
101
- "local_extracted_archives_paths": local_extracted_audio_paths["train"],
102
- "metadata_paths": hash_meta_paths["train"],
103
- }
104
- ),
105
- ]
106
-
107
- def _generate_examples(self, audio_archives, local_extracted_archives_paths, metadata_paths):
108
-
109
- features = ["speaker_id","gender","duration","normalized_text"]
110
-
111
- with open(metadata_paths) as f:
112
- metadata = {x["audio_id"]: x for x in csv.DictReader(f, delimiter="\t")}
113
-
114
- for audio_archive, local_extracted_archive_path in zip(audio_archives, local_extracted_archives_paths):
115
- for audio_filename, audio_file in audio_archive:
116
- audio_id = audio_filename.split(os.sep)[-1].split(_AUDIO_EXTENSIONS)[0]
117
- path = os.path.join(local_extracted_archive_path, audio_filename) if local_extracted_archive_path else audio_filename
118
-
119
- yield audio_id, {
120
- "audio_id": audio_id,
121
- **{feature: metadata[audio_id][feature] for feature in features},
122
- "audio": {"path": path, "bytes": audio_file.read()},
123
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
corpus/speech/train.tar.gz → ciempiess_balance/train-00000-of-00003.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83380e26879435b313ec9d197cf9d4d6f7ee59e016e13da81b85c28a6b878f71
3
- size 1295843169
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce15ebda86cd7483d88bc0b864874d17b16c102c81e3a0d7abdd9a2a4954ade6
3
+ size 392456833
ciempiess_balance/train-00001-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:211aad4adc4d96e458fff376c2888d7f6055a377b0d34fd8b9fef37b753d1cf3
3
+ size 440405225
ciempiess_balance/train-00002-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d81a5057a442ddbfa074fa86aa93dea92f3b8bc72c408e3bc2cf1853b659aab1
3
+ size 462574467
corpus/files/metadata_train.tsv DELETED
The diff for this file is too large to render. See raw diff
 
corpus/files/tars_train.paths DELETED
@@ -1 +0,0 @@
1
- corpus/speech/train.tar.gz