|
|
|
|
|
from pathlib import Path |
|
from typing import Dict, List, Tuple |
|
|
|
import datasets |
|
import jsonlines |
|
|
|
from seacrowd.utils import schemas |
|
from seacrowd.utils.configs import SEACrowdConfig |
|
from seacrowd.utils.constants import Licenses, Tasks |
|
|
|
_CITATION = """\ |
|
@inproceedings{huguet-cabot-et-al-2023-redfm-dataset, |
|
title = "RED$^{\rm FM}$: a Filtered and Multilingual Relation Extraction Dataset", |
|
author = "Huguet Cabot, Pere-Lluís and Tedeschi, Simone and Ngonga Ngomo, Axel-Cyrille and |
|
Navigli, Roberto", |
|
booktitle = "Proc. of the 61st Annual Meeting of the Association for Computational Linguistics: ACL 2023", |
|
month = jul, |
|
year = "2023", |
|
address = "Toronto, Canada", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://arxiv.org/abs/2306.09802", |
|
} |
|
""" |
|
|
|
_DATASETNAME = "sredfm" |
|
|
|
|
|
_DESCRIPTION = """\ |
|
SREDFM is an automatically annotated dataset for relation extraction task covering 18 languages, 400 relation types, 13 entity types, totaling more than 40 million triplet instances. SREDFM includes Vietnamnese. |
|
""" |
|
|
|
_HOMEPAGE = "https://github.com/babelscape/rebel" |
|
|
|
_LANGUAGES = ["vie"] |
|
|
|
_LICENSE = Licenses.CC_BY_SA_4_0.value |
|
|
|
_LOCAL = False |
|
|
|
_URLS = { |
|
"train": "https://huggingface.co/datasets/Babelscape/SREDFM/resolve/main/data/train.vi.jsonl", |
|
"dev": "https://huggingface.co/datasets/Babelscape/SREDFM/resolve/main/data/dev.vi.jsonl", |
|
"test": "https://huggingface.co/datasets/Babelscape/SREDFM/resolve/main/data/test.vi.jsonl", |
|
"relations_url": "https://huggingface.co/datasets/Babelscape/SREDFM/raw/main/relations.tsv", |
|
} |
|
|
|
_SUPPORTED_TASKS = [Tasks.RELATION_EXTRACTION] |
|
|
|
_SOURCE_VERSION = "1.0.0" |
|
|
|
_SEACROWD_VERSION = "2024.06.20" |
|
|
|
|
|
class SREDFMDataset(datasets.GeneratorBasedBuilder): |
|
"""SREDFM is an automatically annotated dataset for relation extraction task. |
|
Relation Extraction (RE) is a task that identifies relationships between entities in a text, |
|
enabling the acquisition of relational facts and bridging the gap between natural language |
|
and structured knowledge. SREDFM covers 400 relation types, 13 entity types, |
|
totaling more than 40 million triplet instances.""" |
|
|
|
SOURCE_VERSION = datasets.Version(_SOURCE_VERSION) |
|
SEACROWD_VERSION = datasets.Version(_SEACROWD_VERSION) |
|
|
|
BUILDER_CONFIGS = [ |
|
SEACrowdConfig( |
|
name=f"{_DATASETNAME}_source", |
|
version=SOURCE_VERSION, |
|
description=f"{_DATASETNAME} source schema", |
|
schema="source", |
|
subset_id=f"{_DATASETNAME}", |
|
), |
|
SEACrowdConfig( |
|
name=f"{_DATASETNAME}_seacrowd_kb", |
|
version=SEACROWD_VERSION, |
|
description=f"{_DATASETNAME} SEACrowd schema", |
|
schema="seacrowd_kb", |
|
subset_id=f"{_DATASETNAME}", |
|
), |
|
] |
|
|
|
DEFAULT_CONFIG_NAME = f"{_DATASETNAME}_source" |
|
|
|
def _info(self) -> datasets.DatasetInfo: |
|
if self.config.schema == "source": |
|
features = datasets.Features( |
|
{ |
|
"docid": datasets.Value("string"), |
|
"title": datasets.Value("string"), |
|
"uri": datasets.Value("string"), |
|
"text": datasets.Value("string"), |
|
"entities": [ |
|
{ |
|
"uri": datasets.Value(dtype="string"), |
|
"surfaceform": datasets.Value(dtype="string"), |
|
"type": datasets.Value(dtype="string"), |
|
"start": datasets.Value(dtype="int32"), |
|
"end": datasets.Value(dtype="int32"), |
|
} |
|
], |
|
"relations": [ |
|
{ |
|
"subject": datasets.Value(dtype="int32"), |
|
"predicate": datasets.Value(dtype="string"), |
|
"object": datasets.Value(dtype="int32"), |
|
} |
|
], |
|
} |
|
) |
|
|
|
elif self.config.schema == "seacrowd_kb": |
|
features = schemas.kb_features |
|
|
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=features, |
|
homepage=_HOMEPAGE, |
|
license=_LICENSE, |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager: datasets.DownloadManager) -> List[datasets.SplitGenerator]: |
|
"""Returns SplitGenerators.""" |
|
data_dir = dl_manager.download_and_extract(_URLS) |
|
|
|
relation_names = dict() |
|
relation_path = data_dir["relations_url"] |
|
with open(relation_path, encoding="utf-8") as f: |
|
for row in f: |
|
rel_code, rel_name, _, _ = row.strip().split("\t") |
|
relation_names[rel_code] = rel_name |
|
|
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={"filepath": data_dir["train"], "relation_names": relation_names}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
gen_kwargs={"filepath": data_dir["test"], "relation_names": relation_names}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.VALIDATION, |
|
gen_kwargs={"filepath": data_dir["dev"], "relation_names": relation_names}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, filepath: Path, relation_names: dict) -> Tuple[int, Dict]: |
|
"""Yields examples as (key, example) tuples.""" |
|
|
|
if self.config.schema == "source": |
|
with jsonlines.open(filepath) as f: |
|
skip = set() |
|
for example in f.iter(): |
|
if example["docid"] in skip: |
|
continue |
|
skip.add(example["docid"]) |
|
|
|
entities = [] |
|
for entity in example["entities"]: |
|
entities.append( |
|
{ |
|
"uri": entity["uri"], |
|
"surfaceform": entity["surfaceform"], |
|
"start": entity["boundaries"][0], |
|
"end": entity["boundaries"][1], |
|
"type": entity["type"], |
|
} |
|
) |
|
|
|
relations = [] |
|
for relation in example["relations"]: |
|
if relation["predicate"]["uri"] not in relation_names or relation["confidence"] <= 0.75: |
|
continue |
|
|
|
relations.append( |
|
{ |
|
"subject": entities.index( |
|
{ |
|
"uri": relation["subject"]["uri"], |
|
"surfaceform": relation["subject"]["surfaceform"], |
|
"start": relation["subject"]["boundaries"][0], |
|
"end": relation["subject"]["boundaries"][1], |
|
"type": relation["subject"]["type"], |
|
} |
|
), |
|
"predicate": relation_names[relation["predicate"]["uri"]], |
|
"object": entities.index( |
|
{ |
|
"uri": relation["object"]["uri"], |
|
"surfaceform": relation["object"]["surfaceform"], |
|
"start": relation["object"]["boundaries"][0], |
|
"end": relation["object"]["boundaries"][1], |
|
"type": relation["object"]["type"], |
|
} |
|
), |
|
} |
|
) |
|
|
|
if len(relations) == 0: |
|
continue |
|
|
|
yield example["docid"], { |
|
"docid": example["docid"], |
|
"title": example["title"], |
|
"uri": example["uri"], |
|
"text": example["text"], |
|
"entities": entities, |
|
"relations": relations, |
|
} |
|
|
|
elif self.config.schema == "seacrowd_kb": |
|
with jsonlines.open(filepath) as f: |
|
skip = set() |
|
i = 0 |
|
for example in f.iter(): |
|
if example["docid"] in skip: |
|
continue |
|
skip.add(example["docid"]) |
|
|
|
i += 1 |
|
processed_text = example["text"].replace("\n", " ") |
|
passages = [ |
|
{ |
|
"id": f"{i}-{example['uri']}", |
|
"type": "text", |
|
"text": [processed_text], |
|
"offsets": [[0, len(processed_text)]], |
|
} |
|
] |
|
|
|
entities = [] |
|
for entity in example["entities"]: |
|
entities.append( |
|
{ |
|
"id": entity["uri"], |
|
"type": entity["type"], |
|
"text": [entity["surfaceform"]], |
|
"offsets": [entity["boundaries"]], |
|
"normalized": {"db_name": "", "db_id": ""}, |
|
} |
|
) |
|
|
|
relations = [] |
|
for relation in example["relations"]: |
|
if relation["predicate"]["uri"] not in relation_names or relation["confidence"] <= 0.75: |
|
continue |
|
|
|
i += 1 |
|
sub = relation["subject"] |
|
pred = relation["predicate"] |
|
obj = relation["object"] |
|
relations.append( |
|
{ |
|
"id": f"{i}-{sub['uri']}-{pred['uri']}-{obj['uri']}", |
|
"type": relation_names[pred["uri"]], |
|
"arg1_id": str( |
|
entities.index( |
|
{ |
|
"id": sub["uri"], |
|
"type": sub["type"], |
|
"text": [sub["surfaceform"]], |
|
"offsets": [sub["boundaries"]], |
|
"normalized": {"db_name": "", "db_id": ""}, |
|
} |
|
) |
|
), |
|
"arg2_id": str( |
|
entities.index( |
|
{ |
|
"id": obj["uri"], |
|
"type": obj["type"], |
|
"text": [obj["surfaceform"]], |
|
"offsets": [obj["boundaries"]], |
|
"normalized": {"db_name": "", "db_id": ""}, |
|
} |
|
) |
|
), |
|
"normalized": {"db_name": "", "db_id": ""}, |
|
} |
|
) |
|
|
|
for entity in entities: |
|
i += 1 |
|
entity["id"] = f"{i}-{entity['id']}" |
|
|
|
if len(relations) == 0: |
|
continue |
|
|
|
yield example["docid"], { |
|
"id": example["docid"], |
|
"passages": passages, |
|
"entities": entities, |
|
"relations": relations, |
|
"events": [], |
|
"coreferences": [], |
|
} |
|
|