--- annotations_creators: - no-annotation language_creators: - crowdsourced language: - ca license: cc-by-nc-4.0 multilinguality: - monolingual size_categories: - 1K and . ## Dataset Details ### Dataset Description The audios from this dataset have been created with professional studio recordings by professional voice actors in Lafresca Creative Studio. We processed the raw audios with the following recipe: - **Trimming:** Long silences from the start and the end of clips have been removed. - [py-webrtcvad](https://pypi.org/project/webrtcvad/) -> Python interface to the Voice Activity Detector (VAD) developed by Google for the WebRTC. - **Resampling:** From 48000 Hz to 22050 Hz, which is the most common sampling rate for training TTS models. - [SOX](https://github.com/chirlu/sox) -> SOX is a versatile command-line utility primarily used for converting audio files between different formats. - **Stereo to mono conversion:** The original raw audios were provided in stereo, but for TTS training the audios neede to be in mono. Because of that, we used the libraries librosa and soundfile to make this conversion and posterior check. - [Librosa](https://librosa.org/) Librosa is a Python library for music and audio analysis, offering tools for feature extraction, audio manipulation, and visualization. - [Soundfile](https://github.com/bastibe/python-soundfile) SoundFile is a library for reading and writing sound files in various formats, providing a simple interface for audio I/O operations. In total, there are 4 different accents, with 2 speakers per accent (female and male). After trimming, accumulates a total of 3,75h (divided by speaker IDs) as follows: * Balear * olga -> 23.5 min * quim -> 30.93 min * Central * elia -> 33.14 min * grau -> 37,86 min * Occidental (North-Western) * emma -> 28,67 min * pere -> 25,12 min * Valencia * gina -> 22,25 min * lluc -> 23,58 min ## Uses The purpose of this dataset is mainly for training text-to-speech and automatic speech recognition models in Catalan dialects. ### Languages The dataset is in Catalan (`ca-ES`). ## Dataset Structure The dataset consists of a single split, providing audios and transcriptions: ``` DatasetDict({ train: Dataset({ features: ['audio', 'transcription', 'speaker_id', 'accent'], num_rows: 2858 }) }) ``` Each data point is structured as: ``` >> audio_dataset[0] {'audio': {'path': 'lafresca_multiaccent/central/grau/grau_220.wav', 'array': array([0., 0., 0., ..., 0., 0., 0.]), 'sampling_rate': 22050}, 'transcription': 'Una mica més amunt, un cop passats els blocs de pisos, hi havia una altra casa de dos o tres pisos en la que plantaven floretes petites blanques.', 'speaker_id': 'grau', 'accent': 'central'} ``` ### Dataset Splits - ```audio (dict)```: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: dataset[0]["audio"] the audio file is automatically decoded and resampled to dataset.features["audio"].sampling_rate. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus, it is important to first query the sample index before the "audio" column, i.e. dataset[0]["audio"] should always be preferred over dataset["audio"][0]. * path (str): The path to the audio file. * array (array): Decoded audio array. * sampling_rate (int): Audio sampling rate. - ```transcription (str)```: The sentence the user was prompted to speak. ## Dataset Creation This dataset has been created by members of the Language Technologies unit from the Life Sciences department of the Barcelona Supercomputing Center, except the valencian sentences which were created with the support of Cenid, the Digital Intelligence Center of the University of Alicante. The voices belong to professional voice actors and they've been recorded in Lafresca Creative Studio. ### Source Data The data presented in this dataset is the source data. #### Data Collection and Processing These are the technical details of the data collection and processing: * Microphone: Austrian Audio oc818 * Preamp: Focusrite ISA Two * Audio Interface: Antelope Orion 32+ * DAW: ProTools 2023.6.0 Processing: * Noise Gate: C1 Gate * Compression BF-76 * De-Esser Renaissance * EQ Maag EQ2 * EQ FabFilter Pro-Q3 * Limiter: L1 Ultramaximizer Here's the information about the speakers: | Dialect | Gender | County | |------------|---------|-----------------| | Central | male | Barcelonès | | Central | female | Barcelonès | | Balear | female | Pla de Mallorca | | Balear | male | Llevant | | Occidental | male | Baix Ebre | | Occidental | female | Baix Ebre | | Valencian | female | Ribera Alta | | Valencian | male | La Plana Baixa | #### Who are the source data producers? The Language Technologies team from the Life Sciences department at the Barcelona Supercomputing Center developed this dataset. It features recordings by professional voice actors made at Lafresca Creative Studio. ### Annotations In order to check whether or not there were any errors in the transcriptions of the audios, we created a Label Studio space. In that space, we manually listened to subset of the dataset, and compared what we heard with the transcription. If the transcription was mistaken, we corrected it. #### Personal and Sensitive Information The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset. ## Bias, Risks, and Limitations Training a Text-to-Speech (TTS) model by fine-tuning with a Catalan speaker who speaks a particular dialect presents significant limitations. Mostly, the challenge is in capturing the full range of variability inherent in that accent. Each dialect has its own unique phonetic, intonational, and prosodic characteristics that can vary greatly even within a single linguistic region. Consequently, a TTS model trained on a narrow dialect sample will struggle to generalize across different accents and sub-dialects, leading to reduced accuracy and naturalness. Additionally, achieving a standard representation is exceedingly difficult because linguistic features can differ markedly not only between dialects but also among individual speakers within the same dialect group. These variations encompass subtle nuances in pronunciation, rhythm, and speech patterns that are challenging to standardize in a model trained on a limited dataset. ### Recommendations ## Citation **APA:** ## Funding This work has been promoted and financed by the Generalitat de Catalunya through the [Aina project](https://projecteaina.cat/), in addition the Valencian sentences have been created within the framework of the NEL-VIVES project 2022/TL22/00215334. ## Dataset Card Contact langtech@bsc.es