|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import os |
|
|
|
import datasets |
|
|
|
|
|
_DESCRIPTION = """\ |
|
Parallel corpora from Web Crawls collected in the ParaCrawl project. |
|
|
|
42 languages, 43 bitexts |
|
total number of files: 59,996 |
|
total number of tokens: 56.11G |
|
total number of sentence fragments: 3.13G |
|
""" |
|
_HOMEPAGE = "http://opus.nlpl.eu/ParaCrawl.php" |
|
_CITATION = r"""\ |
|
@inproceedings{banon-etal-2020-paracrawl, |
|
title = "{P}ara{C}rawl: Web-Scale Acquisition of Parallel Corpora", |
|
author = "Ba{\~n}{\'o}n, Marta and |
|
Chen, Pinzhen and |
|
Haddow, Barry and |
|
Heafield, Kenneth and |
|
Hoang, Hieu and |
|
Espl{\`a}-Gomis, Miquel and |
|
Forcada, Mikel L. and |
|
Kamran, Amir and |
|
Kirefu, Faheem and |
|
Koehn, Philipp and |
|
Ortiz Rojas, Sergio and |
|
Pla Sempere, Leopoldo and |
|
Ram{\'\i}rez-S{\'a}nchez, Gema and |
|
Sarr{\'\i}as, Elsa and |
|
Strelec, Marek and |
|
Thompson, Brian and |
|
Waites, William and |
|
Wiggins, Dion and |
|
Zaragoza, Jaume", |
|
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics", |
|
month = jul, |
|
year = "2020", |
|
address = "Online", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://aclanthology.org/2020.acl-main.417", |
|
doi = "10.18653/v1/2020.acl-main.417", |
|
pages = "4555--4567", |
|
} |
|
@InProceedings{TIEDEMANN12.463, |
|
author = {Jörg Tiedemann}, |
|
title = {Parallel Data, Tools and Interfaces in OPUS}, |
|
booktitle = {Proceedings of the Eight International Conference on Language Resources and Evaluation (LREC'12)}, |
|
year = {2012}, |
|
month = {may}, |
|
date = {23-25}, |
|
address = {Istanbul, Turkey}, |
|
editor = {Nicoletta Calzolari (Conference Chair) and Khalid Choukri and Thierry Declerck and Mehmet Uğur Doğan and Bente Maegaard and Joseph Mariani and Asuncion Moreno and Jan Odijk and Stelios Piperidis}, |
|
publisher = {European Language Resources Association (ELRA)}, |
|
isbn = {978-2-9517408-7-7}, |
|
language = {english} |
|
} |
|
""" |
|
|
|
_VERSION = "9.0.0" |
|
_BASE_NAME = "ParaCrawl.{}.{}" |
|
_BASE_URL = "https://object.pouta.csc.fi/OPUS-ParaCrawl/v9/moses/{}-{}.txt.zip" |
|
|
|
_LANGUAGE_PAIRS = [ |
|
("el", "en"), |
|
("en", "km"), |
|
("en", "so"), |
|
("de", "pl"), |
|
("fr", "nl"), |
|
("en", "sw"), |
|
("en", "tl"), |
|
("es", "gl"), |
|
] |
|
|
|
|
|
class ParaCrawlConfig(datasets.BuilderConfig): |
|
def __init__(self, *args, lang1=None, lang2=None, **kwargs): |
|
super().__init__( |
|
*args, |
|
name=f"{lang1}-{lang2}", |
|
**kwargs, |
|
) |
|
assert lang1 != lang2, "'language 1' & 'language 2' should be different from each other" |
|
self.lang1 = lang1 |
|
self.lang2 = lang2 |
|
|
|
|
|
class OpusParaCrawl(datasets.GeneratorBasedBuilder): |
|
BUILDER_CONFIGS = [ |
|
ParaCrawlConfig( |
|
lang1=lang1, |
|
lang2=lang2, |
|
description=f"Translating {lang1} to {lang2} or vice versa", |
|
version=datasets.Version(_VERSION), |
|
) |
|
for lang1, lang2 in _LANGUAGE_PAIRS |
|
] |
|
BUILDER_CONFIG_CLASS = ParaCrawlConfig |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=datasets.Features( |
|
{ |
|
"id": datasets.Value("string"), |
|
"translation": datasets.Translation(languages=(self.config.lang1, self.config.lang2)), |
|
}, |
|
), |
|
supervised_keys=None, |
|
homepage=_HOMEPAGE, |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
download_url = _BASE_URL.format(self.config.lang1, self.config.lang2) |
|
path = dl_manager.download_and_extract(download_url) |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={"datapath": path}, |
|
) |
|
] |
|
|
|
def _generate_examples(self, datapath): |
|
lang1, lang2 = self.config.lang1, self.config.lang2 |
|
folder = lang1 + "-" + lang2 |
|
lang1_filename = _BASE_NAME.format(folder, lang1) |
|
lang2_filename = _BASE_NAME.format(folder, lang2) |
|
lang1_path = os.path.join(datapath, lang1_filename) |
|
lang2_path = os.path.join(datapath, lang2_filename) |
|
with open(lang1_path, encoding="utf-8") as f1, open(lang2_path, encoding="utf-8") as f2: |
|
for id_, (lang1_sentence, lang2_sentence) in enumerate(zip(f1, f2)): |
|
lang1_sentence = lang1_sentence.strip() |
|
lang2_sentence = lang2_sentence.strip() |
|
yield id_, { |
|
"id": str(id_), |
|
"translation": {lang1: lang1_sentence, lang2: lang2_sentence}, |
|
} |
|
|