sha
stringlengths
40
40
text
stringlengths
1
13.4M
id
stringlengths
2
117
tags
sequencelengths
1
7.91k
created_at
stringlengths
25
25
metadata
stringlengths
2
875k
last_modified
stringlengths
25
25
arxiv
sequencelengths
0
25
languages
sequencelengths
0
7.91k
tags_str
stringlengths
17
159k
text_str
stringlengths
1
447k
text_lists
sequencelengths
0
352
processed_texts
sequencelengths
1
353
15ef643450d589d5883e289ffadeb03563e80a9e
# Dataset Card for Acronym Identification Dataset ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://sites.google.com/view/sdu-aaai21/shared-task - **Repository:** https://github.com/amirveyseh/AAAI-21-SDU-shared-task-1-AI - **Paper:** [What Does This Acronym Mean? Introducing a New Dataset for Acronym Identification and Disambiguation](https://arxiv.org/pdf/2010.14678v1.pdf) - **Leaderboard:** https://competitions.codalab.org/competitions/26609 - **Point of Contact:** [More Information Needed] ### Dataset Summary This dataset contains the training, validation, and test data for the **Shared Task 1: Acronym Identification** of the AAAI-21 Workshop on Scientific Document Understanding. ### Supported Tasks and Leaderboards The dataset supports an `acronym-identification` task, where the aim is to predic which tokens in a pre-tokenized sentence correspond to acronyms. The dataset was released for a Shared Task which supported a [leaderboard](https://competitions.codalab.org/competitions/26609). ### Languages The sentences in the dataset are in English (`en`). ## Dataset Structure ### Data Instances A sample from the training set is provided below: ``` {'id': 'TR-0', 'labels': [4, 4, 4, 4, 0, 2, 2, 4, 1, 4, 4, 4, 4, 4, 4, 4, 4, 4], 'tokens': ['What', 'is', 'here', 'called', 'controlled', 'natural', 'language', '(', 'CNL', ')', 'has', 'traditionally', 'been', 'given', 'many', 'different', 'names', '.']} ``` Please note that in test set sentences only the `id` and `tokens` fields are available. `labels` can be ignored for test set. Labels in the test set are all `O` ### Data Fields The data instances have the following fields: - `id`: a `string` variable representing the example id, unique across the full dataset - `tokens`: a list of `string` variables representing the word-tokenized sentence - `labels`: a list of `categorical` variables with possible values `["B-long", "B-short", "I-long", "I-short", "O"]` corresponding to a BIO scheme. `-long` corresponds to the expanded acronym, such as *controlled natural language* here, and `-short` to the abbrviation, `CNL` here. ### Data Splits The training, validation, and test set contain `14,006`, `1,717`, and `1750` sentences respectively. ## Dataset Creation ### Curation Rationale > First, most of the existing datasets for acronym identification (AI) are either limited in their sizes or created using simple rule-based methods. > This is unfortunate as rules are in general not able to capture all the diverse forms to express acronyms and their long forms in text. > Second, most of the existing datasets are in the medical domain, ignoring the challenges in other scientific domains. > In order to address these limitations this paper introduces two new datasets for Acronym Identification. > Notably, our datasets are annotated by human to achieve high quality and have substantially larger numbers of examples than the existing AI datasets in the non-medical domain. ### Source Data #### Initial Data Collection and Normalization > In order to prepare a corpus for acronym annotation, we collect a corpus of 6,786 English papers from arXiv. > These papers consist of 2,031,592 sentences that would be used for data annotation for AI in this work. The dataset paper does not report the exact tokenization method. #### Who are the source language producers? The language was comes from papers hosted on the online digital archive [arXiv](https://arxiv.org/). No more information is available on the selection process or identity of the writers. ### Annotations #### Annotation process > Each sentence for annotation needs to contain at least one word in which more than half of the characters in are capital letters (i.e., acronym candidates). > Afterward, we search for a sub-sequence of words in which the concatenation of the first one, two or three characters of the words (in the order of the words in the sub-sequence could form an acronym candidate. > We call the sub-sequence a long form candidate. If we cannot find any long form candidate, we remove the sentence. > Using this process, we end up with 17,506 sentences to be annotated manually by the annotators from Amazon Mechanical Turk (MTurk). > In particular, we create a HIT for each sentence and ask the workers to annotate the short forms and the long forms in the sentence. > In case of disagreements, if two out of three workers agree on an annotation, we use majority voting to decide the correct annotation. > Otherwise, a fourth annotator is hired to resolve the conflict #### Who are the annotators? Workers were recruited through Amazon MEchanical Turk and paid $0.05 per annotation. No further demographic information is provided. ### Personal and Sensitive Information Papers published on arXiv are unlikely to contain much personal information, although some do include some poorly chosen examples revealing personal details, so the data should be used with care. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information The dataset provided for this shared task is licensed under CC BY-NC-SA 4.0 international license. ### Citation Information ``` @inproceedings{Veyseh2020, author = {Amir Pouran Ben Veyseh and Franck Dernoncourt and Quan Hung Tran and Thien Huu Nguyen}, editor = {Donia Scott and N{\'{u}}ria Bel and Chengqing Zong}, title = {What Does This Acronym Mean? Introducing a New Dataset for Acronym Identification and Disambiguation}, booktitle = {Proceedings of the 28th International Conference on Computational Linguistics, {COLING} 2020, Barcelona, Spain (Online), December 8-13, 2020}, pages = {3285--3301}, publisher = {International Committee on Computational Linguistics}, year = {2020}, url = {https://doi.org/10.18653/v1/2020.coling-main.292}, doi = {10.18653/v1/2020.coling-main.292} } ``` ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
acronym_identification
[ "task_categories:token-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:mit", "acronym-identification", "arxiv:2010.14678", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["en"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": [], "paperswithcode_id": "acronym-identification", "pretty_name": "Acronym Identification Dataset", "tags": ["acronym-identification"], "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "labels", "sequence": {"class_label": {"names": {"0": "B-long", "1": "B-short", "2": "I-long", "3": "I-short", "4": "O"}}}}], "splits": [{"name": "train", "num_bytes": 7792771, "num_examples": 14006}, {"name": "validation", "num_bytes": 952689, "num_examples": 1717}, {"name": "test", "num_bytes": 987712, "num_examples": 1750}], "download_size": 2071007, "dataset_size": 9733172}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}, {"split": "test", "path": "data/test-*"}]}], "train-eval-index": [{"config": "default", "task": "token-classification", "task_id": "entity_extraction", "splits": {"eval_split": "test"}, "col_mapping": {"tokens": "tokens", "labels": "tags"}}]}
2024-01-09T11:39:57+00:00
[ "2010.14678" ]
[ "en" ]
TAGS #task_categories-token-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-mit #acronym-identification #arxiv-2010.14678 #region-us
# Dataset Card for Acronym Identification Dataset ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: URL - Paper: What Does This Acronym Mean? Introducing a New Dataset for Acronym Identification and Disambiguation - Leaderboard: URL - Point of Contact: ### Dataset Summary This dataset contains the training, validation, and test data for the Shared Task 1: Acronym Identification of the AAAI-21 Workshop on Scientific Document Understanding. ### Supported Tasks and Leaderboards The dataset supports an 'acronym-identification' task, where the aim is to predic which tokens in a pre-tokenized sentence correspond to acronyms. The dataset was released for a Shared Task which supported a leaderboard. ### Languages The sentences in the dataset are in English ('en'). ## Dataset Structure ### Data Instances A sample from the training set is provided below: Please note that in test set sentences only the 'id' and 'tokens' fields are available. 'labels' can be ignored for test set. Labels in the test set are all 'O' ### Data Fields The data instances have the following fields: - 'id': a 'string' variable representing the example id, unique across the full dataset - 'tokens': a list of 'string' variables representing the word-tokenized sentence - 'labels': a list of 'categorical' variables with possible values '["B-long", "B-short", "I-long", "I-short", "O"]' corresponding to a BIO scheme. '-long' corresponds to the expanded acronym, such as *controlled natural language* here, and '-short' to the abbrviation, 'CNL' here. ### Data Splits The training, validation, and test set contain '14,006', '1,717', and '1750' sentences respectively. ## Dataset Creation ### Curation Rationale > First, most of the existing datasets for acronym identification (AI) are either limited in their sizes or created using simple rule-based methods. > This is unfortunate as rules are in general not able to capture all the diverse forms to express acronyms and their long forms in text. > Second, most of the existing datasets are in the medical domain, ignoring the challenges in other scientific domains. > In order to address these limitations this paper introduces two new datasets for Acronym Identification. > Notably, our datasets are annotated by human to achieve high quality and have substantially larger numbers of examples than the existing AI datasets in the non-medical domain. ### Source Data #### Initial Data Collection and Normalization > In order to prepare a corpus for acronym annotation, we collect a corpus of 6,786 English papers from arXiv. > These papers consist of 2,031,592 sentences that would be used for data annotation for AI in this work. The dataset paper does not report the exact tokenization method. #### Who are the source language producers? The language was comes from papers hosted on the online digital archive arXiv. No more information is available on the selection process or identity of the writers. ### Annotations #### Annotation process > Each sentence for annotation needs to contain at least one word in which more than half of the characters in are capital letters (i.e., acronym candidates). > Afterward, we search for a sub-sequence of words in which the concatenation of the first one, two or three characters of the words (in the order of the words in the sub-sequence could form an acronym candidate. > We call the sub-sequence a long form candidate. If we cannot find any long form candidate, we remove the sentence. > Using this process, we end up with 17,506 sentences to be annotated manually by the annotators from Amazon Mechanical Turk (MTurk). > In particular, we create a HIT for each sentence and ask the workers to annotate the short forms and the long forms in the sentence. > In case of disagreements, if two out of three workers agree on an annotation, we use majority voting to decide the correct annotation. > Otherwise, a fourth annotator is hired to resolve the conflict #### Who are the annotators? Workers were recruited through Amazon MEchanical Turk and paid $0.05 per annotation. No further demographic information is provided. ### Personal and Sensitive Information Papers published on arXiv are unlikely to contain much personal information, although some do include some poorly chosen examples revealing personal details, so the data should be used with care. ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. ## Additional Information ### Dataset Curators ### Licensing Information The dataset provided for this shared task is licensed under CC BY-NC-SA 4.0 international license. ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for Acronym Identification Dataset", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: What Does This Acronym Mean? Introducing a New Dataset for Acronym Identification and Disambiguation\n- Leaderboard: URL\n- Point of Contact:", "### Dataset Summary\n\nThis dataset contains the training, validation, and test data for the Shared Task 1: Acronym Identification of the AAAI-21 Workshop on Scientific Document Understanding.", "### Supported Tasks and Leaderboards\n\nThe dataset supports an 'acronym-identification' task, where the aim is to predic which tokens in a pre-tokenized sentence correspond to acronyms. The dataset was released for a Shared Task which supported a leaderboard.", "### Languages\n\nThe sentences in the dataset are in English ('en').", "## Dataset Structure", "### Data Instances\n\nA sample from the training set is provided below:\n\n\n\nPlease note that in test set sentences only the 'id' and 'tokens' fields are available. 'labels' can be ignored for test set. Labels in the test set are all 'O'", "### Data Fields\n\nThe data instances have the following fields:\n\n- 'id': a 'string' variable representing the example id, unique across the full dataset\n- 'tokens': a list of 'string' variables representing the word-tokenized sentence\n- 'labels': a list of 'categorical' variables with possible values '[\"B-long\", \"B-short\", \"I-long\", \"I-short\", \"O\"]' corresponding to a BIO scheme. '-long' corresponds to the expanded acronym, such as *controlled natural language* here, and '-short' to the abbrviation, 'CNL' here.", "### Data Splits\n\nThe training, validation, and test set contain '14,006', '1,717', and '1750' sentences respectively.", "## Dataset Creation", "### Curation Rationale\n\n> First, most of the existing datasets for acronym identification (AI) are either limited in their sizes or created using simple rule-based methods.\n> This is unfortunate as rules are in general not able to capture all the diverse forms to express acronyms and their long forms in text.\n> Second, most of the existing datasets are in the medical domain, ignoring the challenges in other scientific domains.\n> In order to address these limitations this paper introduces two new datasets for Acronym Identification.\n> Notably, our datasets are annotated by human to achieve high quality and have substantially larger numbers of examples than the existing AI datasets in the non-medical domain.", "### Source Data", "#### Initial Data Collection and Normalization\n\n> In order to prepare a corpus for acronym annotation, we collect a corpus of 6,786 English papers from arXiv.\n> These papers consist of 2,031,592 sentences that would be used for data annotation for AI in this work.\n\nThe dataset paper does not report the exact tokenization method.", "#### Who are the source language producers?\n\nThe language was comes from papers hosted on the online digital archive arXiv. No more information is available on the selection process or identity of the writers.", "### Annotations", "#### Annotation process\n\n> Each sentence for annotation needs to contain at least one word in which more than half of the characters in are capital letters (i.e., acronym candidates).\n> Afterward, we search for a sub-sequence of words in which the concatenation of the first one, two or three characters of the words (in the order of the words in the sub-sequence could form an acronym candidate.\n> We call the sub-sequence a long form candidate. If we cannot find any long form candidate, we remove the sentence.\n> Using this process, we end up with 17,506 sentences to be annotated manually by the annotators from Amazon Mechanical Turk (MTurk).\n> In particular, we create a HIT for each sentence and ask the workers to annotate the short forms and the long forms in the sentence.\n> In case of disagreements, if two out of three workers agree on an annotation, we use majority voting to decide the correct annotation.\n> Otherwise, a fourth annotator is hired to resolve the conflict", "#### Who are the annotators?\n\nWorkers were recruited through Amazon MEchanical Turk and paid $0.05 per annotation. No further demographic information is provided.", "### Personal and Sensitive Information\n\nPapers published on arXiv are unlikely to contain much personal information, although some do include some poorly chosen examples revealing personal details, so the data should be used with care.", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\nDataset provided for research purposes only. Please check dataset license for additional information.", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nThe dataset provided for this shared task is licensed under CC BY-NC-SA 4.0 international license.", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-mit #acronym-identification #arxiv-2010.14678 #region-us \n", "# Dataset Card for Acronym Identification Dataset", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: What Does This Acronym Mean? Introducing a New Dataset for Acronym Identification and Disambiguation\n- Leaderboard: URL\n- Point of Contact:", "### Dataset Summary\n\nThis dataset contains the training, validation, and test data for the Shared Task 1: Acronym Identification of the AAAI-21 Workshop on Scientific Document Understanding.", "### Supported Tasks and Leaderboards\n\nThe dataset supports an 'acronym-identification' task, where the aim is to predic which tokens in a pre-tokenized sentence correspond to acronyms. The dataset was released for a Shared Task which supported a leaderboard.", "### Languages\n\nThe sentences in the dataset are in English ('en').", "## Dataset Structure", "### Data Instances\n\nA sample from the training set is provided below:\n\n\n\nPlease note that in test set sentences only the 'id' and 'tokens' fields are available. 'labels' can be ignored for test set. Labels in the test set are all 'O'", "### Data Fields\n\nThe data instances have the following fields:\n\n- 'id': a 'string' variable representing the example id, unique across the full dataset\n- 'tokens': a list of 'string' variables representing the word-tokenized sentence\n- 'labels': a list of 'categorical' variables with possible values '[\"B-long\", \"B-short\", \"I-long\", \"I-short\", \"O\"]' corresponding to a BIO scheme. '-long' corresponds to the expanded acronym, such as *controlled natural language* here, and '-short' to the abbrviation, 'CNL' here.", "### Data Splits\n\nThe training, validation, and test set contain '14,006', '1,717', and '1750' sentences respectively.", "## Dataset Creation", "### Curation Rationale\n\n> First, most of the existing datasets for acronym identification (AI) are either limited in their sizes or created using simple rule-based methods.\n> This is unfortunate as rules are in general not able to capture all the diverse forms to express acronyms and their long forms in text.\n> Second, most of the existing datasets are in the medical domain, ignoring the challenges in other scientific domains.\n> In order to address these limitations this paper introduces two new datasets for Acronym Identification.\n> Notably, our datasets are annotated by human to achieve high quality and have substantially larger numbers of examples than the existing AI datasets in the non-medical domain.", "### Source Data", "#### Initial Data Collection and Normalization\n\n> In order to prepare a corpus for acronym annotation, we collect a corpus of 6,786 English papers from arXiv.\n> These papers consist of 2,031,592 sentences that would be used for data annotation for AI in this work.\n\nThe dataset paper does not report the exact tokenization method.", "#### Who are the source language producers?\n\nThe language was comes from papers hosted on the online digital archive arXiv. No more information is available on the selection process or identity of the writers.", "### Annotations", "#### Annotation process\n\n> Each sentence for annotation needs to contain at least one word in which more than half of the characters in are capital letters (i.e., acronym candidates).\n> Afterward, we search for a sub-sequence of words in which the concatenation of the first one, two or three characters of the words (in the order of the words in the sub-sequence could form an acronym candidate.\n> We call the sub-sequence a long form candidate. If we cannot find any long form candidate, we remove the sentence.\n> Using this process, we end up with 17,506 sentences to be annotated manually by the annotators from Amazon Mechanical Turk (MTurk).\n> In particular, we create a HIT for each sentence and ask the workers to annotate the short forms and the long forms in the sentence.\n> In case of disagreements, if two out of three workers agree on an annotation, we use majority voting to decide the correct annotation.\n> Otherwise, a fourth annotator is hired to resolve the conflict", "#### Who are the annotators?\n\nWorkers were recruited through Amazon MEchanical Turk and paid $0.05 per annotation. No further demographic information is provided.", "### Personal and Sensitive Information\n\nPapers published on arXiv are unlikely to contain much personal information, although some do include some poorly chosen examples revealing personal details, so the data should be used with care.", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\nDataset provided for research purposes only. Please check dataset license for additional information.", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nThe dataset provided for this shared task is licensed under CC BY-NC-SA 4.0 international license.", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
4ba01c71687dd7c996597042449448ea312126cf
# Dataset Card for Adverse Drug Reaction Data v2 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.sciencedirect.com/science/article/pii/S1532046412000615 - **Repository:** [Needs More Information] - **Paper:** https://www.sciencedirect.com/science/article/pii/S1532046412000615 - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Needs More Information] ### Dataset Summary ADE-Corpus-V2 Dataset: Adverse Drug Reaction Data. This is a dataset for Classification if a sentence is ADE-related (True) or not (False) and Relation Extraction between Adverse Drug Event and Drug. DRUG-AE.rel provides relations between drugs and adverse effects. DRUG-DOSE.rel provides relations between drugs and dosages. ADE-NEG.txt provides all sentences in the ADE corpus that DO NOT contain any drug-related adverse effects. ### Supported Tasks and Leaderboards Sentiment classification, Relation Extraction ### Languages English ## Dataset Structure ### Data Instances #### Config - `Ade_corpus_v2_classification` ``` { 'label': 1, 'text': 'Intravenous azithromycin-induced ototoxicity.' } ``` #### Config - `Ade_corpus_v2_drug_ade_relation` ``` { 'drug': 'azithromycin', 'effect': 'ototoxicity', 'indexes': { 'drug': { 'end_char': [24], 'start_char': [12] }, 'effect': { 'end_char': [44], 'start_char': [33] } }, 'text': 'Intravenous azithromycin-induced ototoxicity.' } ``` #### Config - `Ade_corpus_v2_drug_dosage_relation` ``` { 'dosage': '4 times per day', 'drug': 'insulin', 'indexes': { 'dosage': { 'end_char': [56], 'start_char': [41] }, 'drug': { 'end_char': [40], 'start_char': [33]} }, 'text': 'She continued to receive regular insulin 4 times per day over the following 3 years with only occasional hives.' } ``` ### Data Fields #### Config - `Ade_corpus_v2_classification` - `text` - Input text. - `label` - Whether the adverse drug effect(ADE) related (1) or not (0). - #### Config - `Ade_corpus_v2_drug_ade_relation` - `text` - Input text. - `drug` - Name of drug. - `effect` - Effect caused by the drug. - `indexes.drug.start_char` - Start index of `drug` string in text. - `indexes.drug.end_char` - End index of `drug` string in text. - `indexes.effect.start_char` - Start index of `effect` string in text. - `indexes.effect.end_char` - End index of `effect` string in text. #### Config - `Ade_corpus_v2_drug_dosage_relation` - `text` - Input text. - `drug` - Name of drug. - `dosage` - Dosage of the drug. - `indexes.drug.start_char` - Start index of `drug` string in text. - `indexes.drug.end_char` - End index of `drug` string in text. - `indexes.dosage.start_char` - Start index of `dosage` string in text. - `indexes.dosage.end_char` - End index of `dosage` string in text. ### Data Splits | Train | | ------ | | 23516 | ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information ``` @article{GURULINGAPPA2012885, title = "Development of a benchmark corpus to support the automatic extraction of drug-related adverse effects from medical case reports", journal = "Journal of Biomedical Informatics", volume = "45", number = "5", pages = "885 - 892", year = "2012", note = "Text Mining and Natural Language Processing in Pharmacogenomics", issn = "1532-0464", doi = "https://doi.org/10.1016/j.jbi.2012.04.008", url = "http://www.sciencedirect.com/science/article/pii/S1532046412000615", author = "Harsha Gurulingappa and Abdul Mateen Rajput and Angus Roberts and Juliane Fluck and Martin Hofmann-Apitius and Luca Toldo", keywords = "Adverse drug effect, Benchmark corpus, Annotation, Harmonization, Sentence classification", abstract = "A significant amount of information about drug-related safety issues such as adverse effects are published in medical case reports that can only be explored by human readers due to their unstructured nature. The work presented here aims at generating a systematically annotated corpus that can support the development and validation of methods for the automatic extraction of drug-related adverse effects from medical case reports. The documents are systematically double annotated in various rounds to ensure consistent annotations. The annotated documents are finally harmonized to generate representative consensus annotations. In order to demonstrate an example use case scenario, the corpus was employed to train and validate models for the classification of informative against the non-informative sentences. A Maximum Entropy classifier trained with simple features and evaluated by 10-fold cross-validation resulted in the F1 score of 0.70 indicating a potential useful application of the corpus." } ``` ### Contributions Thanks to [@Nilanshrajput](https://github.com/Nilanshrajput), [@lhoestq](https://github.com/lhoestq) for adding this dataset.
ade_corpus_v2
[ "task_categories:text-classification", "task_categories:token-classification", "task_ids:coreference-resolution", "task_ids:fact-checking", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "size_categories:1K<n<10K", "size_categories:n<1K", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K", "1K<n<10K", "n<1K"], "source_datasets": ["original"], "task_categories": ["text-classification", "token-classification"], "task_ids": ["coreference-resolution", "fact-checking"], "pretty_name": "Adverse Drug Reaction Data v2", "config_names": ["Ade_corpus_v2_classification", "Ade_corpus_v2_drug_ade_relation", "Ade_corpus_v2_drug_dosage_relation"], "dataset_info": [{"config_name": "Ade_corpus_v2_classification", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "Not-Related", "1": "Related"}}}}], "splits": [{"name": "train", "num_bytes": 3403699, "num_examples": 23516}], "download_size": 1706476, "dataset_size": 3403699}, {"config_name": "Ade_corpus_v2_drug_ade_relation", "features": [{"name": "text", "dtype": "string"}, {"name": "drug", "dtype": "string"}, {"name": "effect", "dtype": "string"}, {"name": "indexes", "struct": [{"name": "drug", "sequence": [{"name": "start_char", "dtype": "int32"}, {"name": "end_char", "dtype": "int32"}]}, {"name": "effect", "sequence": [{"name": "start_char", "dtype": "int32"}, {"name": "end_char", "dtype": "int32"}]}]}], "splits": [{"name": "train", "num_bytes": 1545993, "num_examples": 6821}], "download_size": 491362, "dataset_size": 1545993}, {"config_name": "Ade_corpus_v2_drug_dosage_relation", "features": [{"name": "text", "dtype": "string"}, {"name": "drug", "dtype": "string"}, {"name": "dosage", "dtype": "string"}, {"name": "indexes", "struct": [{"name": "drug", "sequence": [{"name": "start_char", "dtype": "int32"}, {"name": "end_char", "dtype": "int32"}]}, {"name": "dosage", "sequence": [{"name": "start_char", "dtype": "int32"}, {"name": "end_char", "dtype": "int32"}]}]}], "splits": [{"name": "train", "num_bytes": 64697, "num_examples": 279}], "download_size": 33004, "dataset_size": 64697}], "configs": [{"config_name": "Ade_corpus_v2_classification", "data_files": [{"split": "train", "path": "Ade_corpus_v2_classification/train-*"}]}, {"config_name": "Ade_corpus_v2_drug_ade_relation", "data_files": [{"split": "train", "path": "Ade_corpus_v2_drug_ade_relation/train-*"}]}, {"config_name": "Ade_corpus_v2_drug_dosage_relation", "data_files": [{"split": "train", "path": "Ade_corpus_v2_drug_dosage_relation/train-*"}]}], "train-eval-index": [{"config": "Ade_corpus_v2_classification", "task": "text-classification", "task_id": "multi_class_classification", "splits": {"train_split": "train"}, "col_mapping": {"text": "text", "label": "target"}, "metrics": [{"type": "accuracy", "name": "Accuracy"}, {"type": "f1", "name": "F1 macro", "args": {"average": "macro"}}, {"type": "f1", "name": "F1 micro", "args": {"average": "micro"}}, {"type": "f1", "name": "F1 weighted", "args": {"average": "weighted"}}, {"type": "precision", "name": "Precision macro", "args": {"average": "macro"}}, {"type": "precision", "name": "Precision micro", "args": {"average": "micro"}}, {"type": "precision", "name": "Precision weighted", "args": {"average": "weighted"}}, {"type": "recall", "name": "Recall macro", "args": {"average": "macro"}}, {"type": "recall", "name": "Recall micro", "args": {"average": "micro"}}, {"type": "recall", "name": "Recall weighted", "args": {"average": "weighted"}}]}]}
2024-01-09T11:42:58+00:00
[]
[ "en" ]
TAGS #task_categories-text-classification #task_categories-token-classification #task_ids-coreference-resolution #task_ids-fact-checking #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-original #language-English #license-unknown #region-us
Dataset Card for Adverse Drug Reaction Data v2 ============================================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: URL * Leaderboard: * Point of Contact: ### Dataset Summary ADE-Corpus-V2 Dataset: Adverse Drug Reaction Data. This is a dataset for Classification if a sentence is ADE-related (True) or not (False) and Relation Extraction between Adverse Drug Event and Drug. URL provides relations between drugs and adverse effects. URL provides relations between drugs and dosages. URL provides all sentences in the ADE corpus that DO NOT contain any drug-related adverse effects. ### Supported Tasks and Leaderboards Sentiment classification, Relation Extraction ### Languages English Dataset Structure ----------------- ### Data Instances #### Config - 'Ade\_corpus\_v2\_classification' #### Config - 'Ade\_corpus\_v2\_drug\_ade\_relation' #### Config - 'Ade\_corpus\_v2\_drug\_dosage\_relation' ### Data Fields #### Config - 'Ade\_corpus\_v2\_classification' * 'text' - Input text. * 'label' - Whether the adverse drug effect(ADE) related (1) or not (0). * #### Config - 'Ade\_corpus\_v2\_drug\_ade\_relation' * 'text' - Input text. * 'drug' - Name of drug. * 'effect' - Effect caused by the drug. * 'URL.start\_char' - Start index of 'drug' string in text. * 'URL.end\_char' - End index of 'drug' string in text. * 'URL.start\_char' - Start index of 'effect' string in text. * 'URL.end\_char' - End index of 'effect' string in text. #### Config - 'Ade\_corpus\_v2\_drug\_dosage\_relation' * 'text' - Input text. * 'drug' - Name of drug. * 'dosage' - Dosage of the drug. * 'URL.start\_char' - Start index of 'drug' string in text. * 'URL.end\_char' - End index of 'drug' string in text. * 'URL.start\_char' - Start index of 'dosage' string in text. * 'URL.end\_char' - End index of 'dosage' string in text. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @Nilanshrajput, @lhoestq for adding this dataset.
[ "### Dataset Summary\n\n\nADE-Corpus-V2 Dataset: Adverse Drug Reaction Data.\nThis is a dataset for Classification if a sentence is ADE-related (True) or not (False) and Relation Extraction between Adverse Drug Event and Drug.\nURL provides relations between drugs and adverse effects.\nURL provides relations between drugs and dosages.\nURL provides all sentences in the ADE corpus that DO NOT contain any drug-related adverse effects.", "### Supported Tasks and Leaderboards\n\n\nSentiment classification, Relation Extraction", "### Languages\n\n\nEnglish\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### Config - 'Ade\\_corpus\\_v2\\_classification'", "#### Config - 'Ade\\_corpus\\_v2\\_drug\\_ade\\_relation'", "#### Config - 'Ade\\_corpus\\_v2\\_drug\\_dosage\\_relation'", "### Data Fields", "#### Config - 'Ade\\_corpus\\_v2\\_classification'\n\n\n* 'text' - Input text.\n* 'label' - Whether the adverse drug effect(ADE) related (1) or not (0).\n*", "#### Config - 'Ade\\_corpus\\_v2\\_drug\\_ade\\_relation'\n\n\n* 'text' - Input text.\n* 'drug' - Name of drug.\n* 'effect' - Effect caused by the drug.\n* 'URL.start\\_char' - Start index of 'drug' string in text.\n* 'URL.end\\_char' - End index of 'drug' string in text.\n* 'URL.start\\_char' - Start index of 'effect' string in text.\n* 'URL.end\\_char' - End index of 'effect' string in text.", "#### Config - 'Ade\\_corpus\\_v2\\_drug\\_dosage\\_relation'\n\n\n* 'text' - Input text.\n* 'drug' - Name of drug.\n* 'dosage' - Dosage of the drug.\n* 'URL.start\\_char' - Start index of 'drug' string in text.\n* 'URL.end\\_char' - End index of 'drug' string in text.\n* 'URL.start\\_char' - Start index of 'dosage' string in text.\n* 'URL.end\\_char' - End index of 'dosage' string in text.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @Nilanshrajput, @lhoestq for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_categories-token-classification #task_ids-coreference-resolution #task_ids-fact-checking #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-original #language-English #license-unknown #region-us \n", "### Dataset Summary\n\n\nADE-Corpus-V2 Dataset: Adverse Drug Reaction Data.\nThis is a dataset for Classification if a sentence is ADE-related (True) or not (False) and Relation Extraction between Adverse Drug Event and Drug.\nURL provides relations between drugs and adverse effects.\nURL provides relations between drugs and dosages.\nURL provides all sentences in the ADE corpus that DO NOT contain any drug-related adverse effects.", "### Supported Tasks and Leaderboards\n\n\nSentiment classification, Relation Extraction", "### Languages\n\n\nEnglish\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### Config - 'Ade\\_corpus\\_v2\\_classification'", "#### Config - 'Ade\\_corpus\\_v2\\_drug\\_ade\\_relation'", "#### Config - 'Ade\\_corpus\\_v2\\_drug\\_dosage\\_relation'", "### Data Fields", "#### Config - 'Ade\\_corpus\\_v2\\_classification'\n\n\n* 'text' - Input text.\n* 'label' - Whether the adverse drug effect(ADE) related (1) or not (0).\n*", "#### Config - 'Ade\\_corpus\\_v2\\_drug\\_ade\\_relation'\n\n\n* 'text' - Input text.\n* 'drug' - Name of drug.\n* 'effect' - Effect caused by the drug.\n* 'URL.start\\_char' - Start index of 'drug' string in text.\n* 'URL.end\\_char' - End index of 'drug' string in text.\n* 'URL.start\\_char' - Start index of 'effect' string in text.\n* 'URL.end\\_char' - End index of 'effect' string in text.", "#### Config - 'Ade\\_corpus\\_v2\\_drug\\_dosage\\_relation'\n\n\n* 'text' - Input text.\n* 'drug' - Name of drug.\n* 'dosage' - Dosage of the drug.\n* 'URL.start\\_char' - Start index of 'drug' string in text.\n* 'URL.end\\_char' - End index of 'drug' string in text.\n* 'URL.start\\_char' - Start index of 'dosage' string in text.\n* 'URL.end\\_char' - End index of 'dosage' string in text.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @Nilanshrajput, @lhoestq for adding this dataset." ]
c2d5f738db1ad21a4126a144dfbb00cb51e0a4a9
# Dataset Card for adversarialQA ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [adversarialQA homepage](https://adversarialqa.github.io/) - **Repository:** [adversarialQA repository](https://github.com/maxbartolo/adversarialQA) - **Paper:** [Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension](https://arxiv.org/abs/2002.00293) - **Leaderboard:** [Dynabench QA Round 1 Leaderboard](https://dynabench.org/tasks/2#overall) - **Point of Contact:** [Max Bartolo](max.bartolo@ucl.ac.uk) ### Dataset Summary We have created three new Reading Comprehension datasets constructed using an adversarial model-in-the-loop. We use three different models; BiDAF (Seo et al., 2016), BERTLarge (Devlin et al., 2018), and RoBERTaLarge (Liu et al., 2019) in the annotation loop and construct three datasets; D(BiDAF), D(BERT), and D(RoBERTa), each with 10,000 training examples, 1,000 validation, and 1,000 test examples. The adversarial human annotation paradigm ensures that these datasets consist of questions that current state-of-the-art models (at least the ones used as adversaries in the annotation loop) find challenging. The three AdversarialQA round 1 datasets provide a training and evaluation resource for such methods. ### Supported Tasks and Leaderboards `extractive-qa`: The dataset can be used to train a model for Extractive Question Answering, which consists in selecting the answer to a question from a passage. Success on this task is typically measured by achieving a high word-overlap [F1 score](https://huggingface.co/metrics/f1). The [RoBERTa-Large](https://huggingface.co/roberta-large) model trained on all the data combined with [SQuAD](https://arxiv.org/abs/1606.05250) currently achieves 64.35% F1. This task has an active leaderboard and is available as round 1 of the QA task on [Dynabench](https://dynabench.org/tasks/2#overall) and ranks models based on F1 score. ### Languages The text in the dataset is in English. The associated BCP-47 code is `en`. ## Dataset Structure ### Data Instances Data is provided in the same format as SQuAD 1.1. An example is shown below: ``` { "data": [ { "title": "Oxygen", "paragraphs": [ { "context": "Among the most important classes of organic compounds that contain oxygen are (where \"R\" is an organic group): alcohols (R-OH); ethers (R-O-R); ketones (R-CO-R); aldehydes (R-CO-H); carboxylic acids (R-COOH); esters (R-COO-R); acid anhydrides (R-CO-O-CO-R); and amides (R-C(O)-NR2). There are many important organic solvents that contain oxygen, including: acetone, methanol, ethanol, isopropanol, furan, THF, diethyl ether, dioxane, ethyl acetate, DMF, DMSO, acetic acid, and formic acid. Acetone ((CH3)2CO) and phenol (C6H5OH) are used as feeder materials in the synthesis of many different substances. Other important organic compounds that contain oxygen are: glycerol, formaldehyde, glutaraldehyde, citric acid, acetic anhydride, and acetamide. Epoxides are ethers in which the oxygen atom is part of a ring of three atoms.", "qas": [ { "id": "22bbe104aa72aa9b511dd53237deb11afa14d6e3", "question": "In addition to having oxygen, what do alcohols, ethers and esters have in common, according to the article?", "answers": [ { "answer_start": 36, "text": "organic compounds" } ] }, { "id": "4240a8e708c703796347a3702cf1463eed05584a", "question": "What letter does the abbreviation for acid anhydrides both begin and end in?", "answers": [ { "answer_start": 244, "text": "R" } ] }, { "id": "0681a0a5ec852ec6920d6a30f7ef65dced493366", "question": "Which of the organic compounds, in the article, contains nitrogen?", "answers": [ { "answer_start": 262, "text": "amides" } ] }, { "id": "2990efe1a56ccf81938fa5e18104f7d3803069fb", "question": "Which of the important classes of organic compounds, in the article, has a number in its abbreviation?", "answers": [ { "answer_start": 262, "text": "amides" } ] } ] } ] } ] } ``` ### Data Fields - title: the title of the Wikipedia page from which the context is sourced - context: the context/passage - id: a string identifier for each question - answers: a list of all provided answers (one per question in our case, but multiple may exist in SQuAD) with an `answer_start` field which is the character index of the start of the answer span, and a `text` field which is the answer text. Note that no answers are provided in the test set. Indeed, this dataset is part of the DynaBench benchmark, for which you can submit your predictions on the [website](https://dynabench.org/tasks/2#1). ### Data Splits The dataset is composed of three different datasets constructed using different models in the loop: BiDAF, BERT-Large, and RoBERTa-Large. Each of these has 10,000 training examples, 1,000 validation examples, and 1,000 test examples for a total of 30,000/3,000/3,000 train/validation/test examples. ## Dataset Creation ### Curation Rationale This dataset was collected to provide a more challenging and diverse Reading Comprehension dataset to state-of-the-art models. ### Source Data #### Initial Data Collection and Normalization The source passages are from Wikipedia and are the same as those used in [SQuAD v1.1](https://arxiv.org/abs/1606.05250). #### Who are the source language producers? The source language produces are Wikipedia editors for the passages, and human annotators on Mechanical Turk for the questions. ### Annotations #### Annotation process The dataset is collected through an adversarial human annotation process which pairs a human annotator and a reading comprehension model in an interactive setting. The human is presented with a passage for which they write a question and highlight the correct answer. The model then tries to answer the question, and, if it fails to answer correctly, the human wins. Otherwise, the human modifies or re-writes their question until the successfully fool the model. #### Who are the annotators? The annotators are from Amazon Mechanical Turk, geographically restricted the the USA, UK and Canada, having previously successfully completed at least 1,000 HITs, and having a HIT approval rate greater than 98%. Crowdworkers undergo intensive training and qualification prior to annotation. ### Personal and Sensitive Information No annotator identifying details are provided. ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset is to help develop better question answering systems. A system that succeeds at the supported task would be able to provide an accurate extractive answer from a short passage. This dataset is to be seen as a test bed for questions which contemporary state-of-the-art models struggle to answer correctly, thus often requiring more complex comprehension abilities than say detecting phrases explicitly mentioned in the passage with high overlap to the question. It should be noted, however, that the the source passages are both domain-restricted and linguistically specific, and that provided questions and answers do not constitute any particular social application. ### Discussion of Biases The dataset may exhibit various biases in terms of the source passage selection, annotated questions and answers, as well as algorithmic biases resulting from the adversarial annotation protocol. ### Other Known Limitations N/a ## Additional Information ### Dataset Curators This dataset was initially created by Max Bartolo, Alastair Roberts, Johannes Welbl, Sebastian Riedel, and Pontus Stenetorp, during work carried out at University College London (UCL). ### Licensing Information This dataset is distributed under [CC BY-SA 3.0](https://creativecommons.org/licenses/by-sa/3.0/). ### Citation Information ``` @article{bartolo2020beat, author = {Bartolo, Max and Roberts, Alastair and Welbl, Johannes and Riedel, Sebastian and Stenetorp, Pontus}, title = {Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension}, journal = {Transactions of the Association for Computational Linguistics}, volume = {8}, number = {}, pages = {662-678}, year = {2020}, doi = {10.1162/tacl\_a\_00338}, URL = { https://doi.org/10.1162/tacl_a_00338 }, eprint = { https://doi.org/10.1162/tacl_a_00338 }, abstract = { Innovations in annotation methodology have been a catalyst for Reading Comprehension (RC) datasets and models. One recent trend to challenge current RC models is to involve a model in the annotation process: Humans create questions adversarially, such that the model fails to answer them correctly. In this work we investigate this annotation methodology and apply it in three different settings, collecting a total of 36,000 samples with progressively stronger models in the annotation loop. This allows us to explore questions such as the reproducibility of the adversarial effect, transfer from data collected with varying model-in-the-loop strengths, and generalization to data collected without a model. We find that training on adversarially collected samples leads to strong generalization to non-adversarially collected datasets, yet with progressive performance deterioration with increasingly stronger models-in-the-loop. Furthermore, we find that stronger models can still learn from datasets collected with substantially weaker models-in-the-loop. When trained on data collected with a BiDAF model in the loop, RoBERTa achieves 39.9F1 on questions that it cannot answer when trained on SQuAD—only marginally lower than when trained on data collected using RoBERTa itself (41.0F1). } } ``` ### Contributions Thanks to [@maxbartolo](https://github.com/maxbartolo) for adding this dataset.
UCLNLP/adversarial_qa
[ "task_categories:question-answering", "task_ids:extractive-qa", "task_ids:open-domain-qa", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "arxiv:2002.00293", "arxiv:1606.05250", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["question-answering"], "task_ids": ["extractive-qa", "open-domain-qa"], "paperswithcode_id": "adversarialqa", "pretty_name": "adversarialQA", "dataset_info": [{"config_name": "adversarialQA", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "context", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "answers", "sequence": [{"name": "text", "dtype": "string"}, {"name": "answer_start", "dtype": "int32"}]}, {"name": "metadata", "struct": [{"name": "split", "dtype": "string"}, {"name": "model_in_the_loop", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 27858686, "num_examples": 30000}, {"name": "validation", "num_bytes": 2757092, "num_examples": 3000}, {"name": "test", "num_bytes": 2919479, "num_examples": 3000}], "download_size": 5301049, "dataset_size": 33535257}, {"config_name": "dbert", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "context", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "answers", "sequence": [{"name": "text", "dtype": "string"}, {"name": "answer_start", "dtype": "int32"}]}, {"name": "metadata", "struct": [{"name": "split", "dtype": "string"}, {"name": "model_in_the_loop", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 9345521, "num_examples": 10000}, {"name": "validation", "num_bytes": 918156, "num_examples": 1000}, {"name": "test", "num_bytes": 971290, "num_examples": 1000}], "download_size": 2689032, "dataset_size": 11234967}, {"config_name": "dbidaf", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "context", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "answers", "sequence": [{"name": "text", "dtype": "string"}, {"name": "answer_start", "dtype": "int32"}]}, {"name": "metadata", "struct": [{"name": "split", "dtype": "string"}, {"name": "model_in_the_loop", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 9282482, "num_examples": 10000}, {"name": "validation", "num_bytes": 917907, "num_examples": 1000}, {"name": "test", "num_bytes": 946947, "num_examples": 1000}], "download_size": 2721341, "dataset_size": 11147336}, {"config_name": "droberta", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "context", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "answers", "sequence": [{"name": "text", "dtype": "string"}, {"name": "answer_start", "dtype": "int32"}]}, {"name": "metadata", "struct": [{"name": "split", "dtype": "string"}, {"name": "model_in_the_loop", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 9270683, "num_examples": 10000}, {"name": "validation", "num_bytes": 925029, "num_examples": 1000}, {"name": "test", "num_bytes": 1005242, "num_examples": 1000}], "download_size": 2815452, "dataset_size": 11200954}], "configs": [{"config_name": "adversarialQA", "data_files": [{"split": "train", "path": "adversarialQA/train-*"}, {"split": "validation", "path": "adversarialQA/validation-*"}, {"split": "test", "path": "adversarialQA/test-*"}]}, {"config_name": "dbert", "data_files": [{"split": "train", "path": "dbert/train-*"}, {"split": "validation", "path": "dbert/validation-*"}, {"split": "test", "path": "dbert/test-*"}]}, {"config_name": "dbidaf", "data_files": [{"split": "train", "path": "dbidaf/train-*"}, {"split": "validation", "path": "dbidaf/validation-*"}, {"split": "test", "path": "dbidaf/test-*"}]}, {"config_name": "droberta", "data_files": [{"split": "train", "path": "droberta/train-*"}, {"split": "validation", "path": "droberta/validation-*"}, {"split": "test", "path": "droberta/test-*"}]}], "train-eval-index": [{"config": "adversarialQA", "task": "question-answering", "task_id": "extractive_question_answering", "splits": {"train_split": "train", "eval_split": "validation"}, "col_mapping": {"question": "question", "context": "context", "answers": {"text": "text", "answer_start": "answer_start"}}, "metrics": [{"type": "squad", "name": "SQuAD"}]}]}
2023-12-21T14:20:00+00:00
[ "2002.00293", "1606.05250" ]
[ "en" ]
TAGS #task_categories-question-answering #task_ids-extractive-qa #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-2002.00293 #arxiv-1606.05250 #region-us
# Dataset Card for adversarialQA ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: adversarialQA homepage - Repository: adversarialQA repository - Paper: Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension - Leaderboard: Dynabench QA Round 1 Leaderboard - Point of Contact: Max Bartolo ### Dataset Summary We have created three new Reading Comprehension datasets constructed using an adversarial model-in-the-loop. We use three different models; BiDAF (Seo et al., 2016), BERTLarge (Devlin et al., 2018), and RoBERTaLarge (Liu et al., 2019) in the annotation loop and construct three datasets; D(BiDAF), D(BERT), and D(RoBERTa), each with 10,000 training examples, 1,000 validation, and 1,000 test examples. The adversarial human annotation paradigm ensures that these datasets consist of questions that current state-of-the-art models (at least the ones used as adversaries in the annotation loop) find challenging. The three AdversarialQA round 1 datasets provide a training and evaluation resource for such methods. ### Supported Tasks and Leaderboards 'extractive-qa': The dataset can be used to train a model for Extractive Question Answering, which consists in selecting the answer to a question from a passage. Success on this task is typically measured by achieving a high word-overlap F1 score. The RoBERTa-Large model trained on all the data combined with SQuAD currently achieves 64.35% F1. This task has an active leaderboard and is available as round 1 of the QA task on Dynabench and ranks models based on F1 score. ### Languages The text in the dataset is in English. The associated BCP-47 code is 'en'. ## Dataset Structure ### Data Instances Data is provided in the same format as SQuAD 1.1. An example is shown below: ### Data Fields - title: the title of the Wikipedia page from which the context is sourced - context: the context/passage - id: a string identifier for each question - answers: a list of all provided answers (one per question in our case, but multiple may exist in SQuAD) with an 'answer_start' field which is the character index of the start of the answer span, and a 'text' field which is the answer text. Note that no answers are provided in the test set. Indeed, this dataset is part of the DynaBench benchmark, for which you can submit your predictions on the website. ### Data Splits The dataset is composed of three different datasets constructed using different models in the loop: BiDAF, BERT-Large, and RoBERTa-Large. Each of these has 10,000 training examples, 1,000 validation examples, and 1,000 test examples for a total of 30,000/3,000/3,000 train/validation/test examples. ## Dataset Creation ### Curation Rationale This dataset was collected to provide a more challenging and diverse Reading Comprehension dataset to state-of-the-art models. ### Source Data #### Initial Data Collection and Normalization The source passages are from Wikipedia and are the same as those used in SQuAD v1.1. #### Who are the source language producers? The source language produces are Wikipedia editors for the passages, and human annotators on Mechanical Turk for the questions. ### Annotations #### Annotation process The dataset is collected through an adversarial human annotation process which pairs a human annotator and a reading comprehension model in an interactive setting. The human is presented with a passage for which they write a question and highlight the correct answer. The model then tries to answer the question, and, if it fails to answer correctly, the human wins. Otherwise, the human modifies or re-writes their question until the successfully fool the model. #### Who are the annotators? The annotators are from Amazon Mechanical Turk, geographically restricted the the USA, UK and Canada, having previously successfully completed at least 1,000 HITs, and having a HIT approval rate greater than 98%. Crowdworkers undergo intensive training and qualification prior to annotation. ### Personal and Sensitive Information No annotator identifying details are provided. ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset is to help develop better question answering systems. A system that succeeds at the supported task would be able to provide an accurate extractive answer from a short passage. This dataset is to be seen as a test bed for questions which contemporary state-of-the-art models struggle to answer correctly, thus often requiring more complex comprehension abilities than say detecting phrases explicitly mentioned in the passage with high overlap to the question. It should be noted, however, that the the source passages are both domain-restricted and linguistically specific, and that provided questions and answers do not constitute any particular social application. ### Discussion of Biases The dataset may exhibit various biases in terms of the source passage selection, annotated questions and answers, as well as algorithmic biases resulting from the adversarial annotation protocol. ### Other Known Limitations N/a ## Additional Information ### Dataset Curators This dataset was initially created by Max Bartolo, Alastair Roberts, Johannes Welbl, Sebastian Riedel, and Pontus Stenetorp, during work carried out at University College London (UCL). ### Licensing Information This dataset is distributed under CC BY-SA 3.0. ### Contributions Thanks to @maxbartolo for adding this dataset.
[ "# Dataset Card for adversarialQA", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: adversarialQA homepage\n- Repository: adversarialQA repository\n- Paper: Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension\n- Leaderboard: Dynabench QA Round 1 Leaderboard\n- Point of Contact: Max Bartolo", "### Dataset Summary\n\nWe have created three new Reading Comprehension datasets constructed using an adversarial model-in-the-loop.\n\nWe use three different models; BiDAF (Seo et al., 2016), BERTLarge (Devlin et al., 2018), and RoBERTaLarge (Liu et al., 2019) in the annotation loop and construct three datasets; D(BiDAF), D(BERT), and D(RoBERTa), each with 10,000 training examples, 1,000 validation, and 1,000 test examples.\n\nThe adversarial human annotation paradigm ensures that these datasets consist of questions that current state-of-the-art models (at least the ones used as adversaries in the annotation loop) find challenging. The three AdversarialQA round 1 datasets provide a training and evaluation resource for such methods.", "### Supported Tasks and Leaderboards\n\n'extractive-qa': The dataset can be used to train a model for Extractive Question Answering, which consists in selecting the answer to a question from a passage. Success on this task is typically measured by achieving a high word-overlap F1 score. The RoBERTa-Large model trained on all the data combined with SQuAD currently achieves 64.35% F1. This task has an active leaderboard and is available as round 1 of the QA task on Dynabench and ranks models based on F1 score.", "### Languages\n\nThe text in the dataset is in English. The associated BCP-47 code is 'en'.", "## Dataset Structure", "### Data Instances\n\nData is provided in the same format as SQuAD 1.1. An example is shown below:", "### Data Fields\n\n- title: the title of the Wikipedia page from which the context is sourced\n- context: the context/passage\n- id: a string identifier for each question\n- answers: a list of all provided answers (one per question in our case, but multiple may exist in SQuAD) with an 'answer_start' field which is the character index of the start of the answer span, and a 'text' field which is the answer text.\n\nNote that no answers are provided in the test set. Indeed, this dataset is part of the DynaBench benchmark, for which you can submit your predictions on the website.", "### Data Splits\n\nThe dataset is composed of three different datasets constructed using different models in the loop: BiDAF, BERT-Large, and RoBERTa-Large. Each of these has 10,000 training examples, 1,000 validation examples, and 1,000 test examples for a total of 30,000/3,000/3,000 train/validation/test examples.", "## Dataset Creation", "### Curation Rationale\n\nThis dataset was collected to provide a more challenging and diverse Reading Comprehension dataset to state-of-the-art models.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe source passages are from Wikipedia and are the same as those used in SQuAD v1.1.", "#### Who are the source language producers?\n\nThe source language produces are Wikipedia editors for the passages, and human annotators on Mechanical Turk for the questions.", "### Annotations", "#### Annotation process\n\nThe dataset is collected through an adversarial human annotation process which pairs a human annotator and a reading comprehension model in an interactive setting. The human is presented with a passage for which they write a question and highlight the correct answer. The model then tries to answer the question, and, if it fails to answer correctly, the human wins. Otherwise, the human modifies or re-writes their question until the successfully fool the model.", "#### Who are the annotators?\n\nThe annotators are from Amazon Mechanical Turk, geographically restricted the the USA, UK and Canada, having previously successfully completed at least 1,000 HITs, and having a HIT approval rate greater than 98%. Crowdworkers undergo intensive training and qualification prior to annotation.", "### Personal and Sensitive Information\n\nNo annotator identifying details are provided.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\nThe purpose of this dataset is to help develop better question answering systems.\n\nA system that succeeds at the supported task would be able to provide an accurate extractive answer from a short passage. This dataset is to be seen as a test bed for questions which contemporary state-of-the-art models struggle to answer correctly, thus often requiring more complex comprehension abilities than say detecting phrases explicitly mentioned in the passage with high overlap to the question.\n\nIt should be noted, however, that the the source passages are both domain-restricted and linguistically specific, and that provided questions and answers do not constitute any particular social application.", "### Discussion of Biases\n\nThe dataset may exhibit various biases in terms of the source passage selection, annotated questions and answers, as well as algorithmic biases resulting from the adversarial annotation protocol.", "### Other Known Limitations\n\nN/a", "## Additional Information", "### Dataset Curators\n\nThis dataset was initially created by Max Bartolo, Alastair Roberts, Johannes Welbl, Sebastian Riedel, and Pontus Stenetorp, during work carried out at University College London (UCL).", "### Licensing Information\n\nThis dataset is distributed under CC BY-SA 3.0.", "### Contributions\n\nThanks to @maxbartolo for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-extractive-qa #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-2002.00293 #arxiv-1606.05250 #region-us \n", "# Dataset Card for adversarialQA", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: adversarialQA homepage\n- Repository: adversarialQA repository\n- Paper: Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension\n- Leaderboard: Dynabench QA Round 1 Leaderboard\n- Point of Contact: Max Bartolo", "### Dataset Summary\n\nWe have created three new Reading Comprehension datasets constructed using an adversarial model-in-the-loop.\n\nWe use three different models; BiDAF (Seo et al., 2016), BERTLarge (Devlin et al., 2018), and RoBERTaLarge (Liu et al., 2019) in the annotation loop and construct three datasets; D(BiDAF), D(BERT), and D(RoBERTa), each with 10,000 training examples, 1,000 validation, and 1,000 test examples.\n\nThe adversarial human annotation paradigm ensures that these datasets consist of questions that current state-of-the-art models (at least the ones used as adversaries in the annotation loop) find challenging. The three AdversarialQA round 1 datasets provide a training and evaluation resource for such methods.", "### Supported Tasks and Leaderboards\n\n'extractive-qa': The dataset can be used to train a model for Extractive Question Answering, which consists in selecting the answer to a question from a passage. Success on this task is typically measured by achieving a high word-overlap F1 score. The RoBERTa-Large model trained on all the data combined with SQuAD currently achieves 64.35% F1. This task has an active leaderboard and is available as round 1 of the QA task on Dynabench and ranks models based on F1 score.", "### Languages\n\nThe text in the dataset is in English. The associated BCP-47 code is 'en'.", "## Dataset Structure", "### Data Instances\n\nData is provided in the same format as SQuAD 1.1. An example is shown below:", "### Data Fields\n\n- title: the title of the Wikipedia page from which the context is sourced\n- context: the context/passage\n- id: a string identifier for each question\n- answers: a list of all provided answers (one per question in our case, but multiple may exist in SQuAD) with an 'answer_start' field which is the character index of the start of the answer span, and a 'text' field which is the answer text.\n\nNote that no answers are provided in the test set. Indeed, this dataset is part of the DynaBench benchmark, for which you can submit your predictions on the website.", "### Data Splits\n\nThe dataset is composed of three different datasets constructed using different models in the loop: BiDAF, BERT-Large, and RoBERTa-Large. Each of these has 10,000 training examples, 1,000 validation examples, and 1,000 test examples for a total of 30,000/3,000/3,000 train/validation/test examples.", "## Dataset Creation", "### Curation Rationale\n\nThis dataset was collected to provide a more challenging and diverse Reading Comprehension dataset to state-of-the-art models.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe source passages are from Wikipedia and are the same as those used in SQuAD v1.1.", "#### Who are the source language producers?\n\nThe source language produces are Wikipedia editors for the passages, and human annotators on Mechanical Turk for the questions.", "### Annotations", "#### Annotation process\n\nThe dataset is collected through an adversarial human annotation process which pairs a human annotator and a reading comprehension model in an interactive setting. The human is presented with a passage for which they write a question and highlight the correct answer. The model then tries to answer the question, and, if it fails to answer correctly, the human wins. Otherwise, the human modifies or re-writes their question until the successfully fool the model.", "#### Who are the annotators?\n\nThe annotators are from Amazon Mechanical Turk, geographically restricted the the USA, UK and Canada, having previously successfully completed at least 1,000 HITs, and having a HIT approval rate greater than 98%. Crowdworkers undergo intensive training and qualification prior to annotation.", "### Personal and Sensitive Information\n\nNo annotator identifying details are provided.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\nThe purpose of this dataset is to help develop better question answering systems.\n\nA system that succeeds at the supported task would be able to provide an accurate extractive answer from a short passage. This dataset is to be seen as a test bed for questions which contemporary state-of-the-art models struggle to answer correctly, thus often requiring more complex comprehension abilities than say detecting phrases explicitly mentioned in the passage with high overlap to the question.\n\nIt should be noted, however, that the the source passages are both domain-restricted and linguistically specific, and that provided questions and answers do not constitute any particular social application.", "### Discussion of Biases\n\nThe dataset may exhibit various biases in terms of the source passage selection, annotated questions and answers, as well as algorithmic biases resulting from the adversarial annotation protocol.", "### Other Known Limitations\n\nN/a", "## Additional Information", "### Dataset Curators\n\nThis dataset was initially created by Max Bartolo, Alastair Roberts, Johannes Welbl, Sebastian Riedel, and Pontus Stenetorp, during work carried out at University College London (UCL).", "### Licensing Information\n\nThis dataset is distributed under CC BY-SA 3.0.", "### Contributions\n\nThanks to @maxbartolo for adding this dataset." ]
2305f2e63b68056f9b9037a3805c8c196e0d5581
# Dataset Card for "aeslc" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** https://github.com/ryanzhumich/AESLC - **Paper:** [This Email Could Save Your Life: Introducing the Task of Email Subject Line Generation](https://arxiv.org/abs/1906.03497) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 11.64 MB - **Size of the generated dataset:** 14.95 MB - **Total amount of disk used:** 26.59 MB ### Dataset Summary A collection of email messages of employees in the Enron Corporation. There are two features: - email_body: email body text. - subject_line: email subject text. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages Monolingual English (mainly en-US) with some exceptions. ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 11.64 MB - **Size of the generated dataset:** 14.95 MB - **Total amount of disk used:** 26.59 MB An example of 'train' looks as follows. ``` { "email_body": "B/C\n<<some doc>>\n", "subject_line": "Service Agreement" } ``` ### Data Fields The data fields are the same among all splits. #### default - `email_body`: a `string` feature. - `subject_line`: a `string` feature. ### Data Splits | name |train|validation|test| |-------|----:|---------:|---:| |default|14436| 1960|1906| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @inproceedings{zhang-tetreault-2019-email, title = "This Email Could Save Your Life: Introducing the Task of Email Subject Line Generation", author = "Zhang, Rui and Tetreault, Joel", booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", month = jul, year = "2019", address = "Florence, Italy", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/P19-1043", doi = "10.18653/v1/P19-1043", pages = "446--456", } ``` ### Contributions Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun) for adding this dataset.
aeslc
[ "task_categories:summarization", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:unknown", "aspect-based-summarization", "conversations-summarization", "multi-document-summarization", "email-headline-generation", "arxiv:1906.03497", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["summarization"], "task_ids": [], "paperswithcode_id": "aeslc", "pretty_name": "AESLC: Annotated Enron Subject Line Corpus", "tags": ["aspect-based-summarization", "conversations-summarization", "multi-document-summarization", "email-headline-generation"], "dataset_info": {"features": [{"name": "email_body", "dtype": "string"}, {"name": "subject_line", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 11897245, "num_examples": 14436}, {"name": "validation", "num_bytes": 1659987, "num_examples": 1960}, {"name": "test", "num_bytes": 1383452, "num_examples": 1906}], "download_size": 7948020, "dataset_size": 14940684}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}, {"split": "test", "path": "data/test-*"}]}]}
2024-01-09T11:49:13+00:00
[ "1906.03497" ]
[ "en" ]
TAGS #task_categories-summarization #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-unknown #aspect-based-summarization #conversations-summarization #multi-document-summarization #email-headline-generation #arxiv-1906.03497 #region-us
Dataset Card for "aeslc" ======================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: * Repository: URL * Paper: This Email Could Save Your Life: Introducing the Task of Email Subject Line Generation * Point of Contact: * Size of downloaded dataset files: 11.64 MB * Size of the generated dataset: 14.95 MB * Total amount of disk used: 26.59 MB ### Dataset Summary A collection of email messages of employees in the Enron Corporation. There are two features: * email\_body: email body text. * subject\_line: email subject text. ### Supported Tasks and Leaderboards ### Languages Monolingual English (mainly en-US) with some exceptions. Dataset Structure ----------------- ### Data Instances #### default * Size of downloaded dataset files: 11.64 MB * Size of the generated dataset: 14.95 MB * Total amount of disk used: 26.59 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### default * 'email\_body': a 'string' feature. * 'subject\_line': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @patrickvonplaten, @thomwolf, @lewtun for adding this dataset.
[ "### Dataset Summary\n\n\nA collection of email messages of employees in the Enron Corporation.\n\n\nThere are two features:\n\n\n* email\\_body: email body text.\n* subject\\_line: email subject text.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nMonolingual English (mainly en-US) with some exceptions.\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 11.64 MB\n* Size of the generated dataset: 14.95 MB\n* Total amount of disk used: 26.59 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'email\\_body': a 'string' feature.\n* 'subject\\_line': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @patrickvonplaten, @thomwolf, @lewtun for adding this dataset." ]
[ "TAGS\n#task_categories-summarization #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-unknown #aspect-based-summarization #conversations-summarization #multi-document-summarization #email-headline-generation #arxiv-1906.03497 #region-us \n", "### Dataset Summary\n\n\nA collection of email messages of employees in the Enron Corporation.\n\n\nThere are two features:\n\n\n* email\\_body: email body text.\n* subject\\_line: email subject text.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nMonolingual English (mainly en-US) with some exceptions.\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 11.64 MB\n* Size of the generated dataset: 14.95 MB\n* Total amount of disk used: 26.59 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'email\\_body': a 'string' feature.\n* 'subject\\_line': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @patrickvonplaten, @thomwolf, @lewtun for adding this dataset." ]
445834a997dce8b40e1d108638064381de80c497
# Dataset Card for Afrikaans Ner Corpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Afrikaans Ner Corpus Homepage](https://repo.sadilar.org/handle/20.500.12185/299) - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** [Martin Puttkammer](mailto:Martin.Puttkammer@nwu.ac.za) ### Dataset Summary The Afrikaans Ner Corpus is an Afrikaans dataset developed by [The Centre for Text Technology (CTexT), North-West University, South Africa](http://humanities.nwu.ac.za/ctext). The data is based on documents from the South African goverment domain and crawled from gov.za websites. It was created to support NER task for Afrikaans language. The dataset uses CoNLL shared task annotation standards. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The language supported is Afrikaans. ## Dataset Structure ### Data Instances A data point consists of sentences seperated by empty line and tab-seperated tokens and tags. {'id': '0', 'ner_tags': [0, 0, 0, 0, 0], 'tokens': ['Vertaling', 'van', 'die', 'inligting', 'in'] } ### Data Fields - `id`: id of the sample - `tokens`: the tokens of the example text - `ner_tags`: the NER tags of each token The NER tags correspond to this list: ``` "OUT", "B-PERS", "I-PERS", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-MISC", "I-MISC", ``` The NER tags have the same format as in the CoNLL shared task: a B denotes the first item of a phrase and an I any non-initial word. There are four types of phrases: person names (PER), organizations (ORG), locations (LOC) and miscellaneous names (MISC). (OUT) is used for tokens not considered part of any named entity. ### Data Splits The data was not split. ## Dataset Creation ### Curation Rationale The data was created to help introduce resources to new language - Afrikaans. [More Information Needed] ### Source Data #### Initial Data Collection and Normalization The data is based on South African government domain and was crawled from gov.za websites. [More Information Needed] #### Who are the source language producers? The data was produced by writers of South African government websites - gov.za [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? The data was annotated during the NCHLT text resource development project. [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators The annotated data sets were developed by the Centre for Text Technology (CTexT, North-West University, South Africa). See: [more information](http://www.nwu.ac.za/ctext) ### Licensing Information The data is under the [Creative Commons Attribution 2.5 South Africa License](http://creativecommons.org/licenses/by/2.5/za/legalcode) ### Citation Information ``` @inproceedings{afrikaans_ner_corpus, author = { Gerhard van Huyssteen and Martin Puttkammer and E.B. Trollip and J.C. Liversage and Roald Eiselen}, title = {NCHLT Afrikaans Named Entity Annotated Corpus}, booktitle = {Eiselen, R. 2016. Government domain named entity recognition for South African languages. Proceedings of the 10th Language Resource and Evaluation Conference, Portorož, Slovenia.}, year = {2016}, url = {https://repo.sadilar.org/handle/20.500.12185/299}, } ``` ### Contributions Thanks to [@yvonnegitau](https://github.com/yvonnegitau) for adding this dataset.
afrikaans_ner_corpus
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:af", "license:other", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["af"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": ["named-entity-recognition"], "pretty_name": "Afrikaans Ner Corpus", "license_details": "Creative Commons Attribution 2.5 South Africa License", "dataset_info": {"config_name": "afrikaans_ner_corpus", "features": [{"name": "id", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "OUT", "1": "B-PERS", "2": "I-PERS", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC", "7": "B-MISC", "8": "I-MISC"}}}}], "splits": [{"name": "train", "num_bytes": 4025651, "num_examples": 8962}], "download_size": 944804, "dataset_size": 4025651}, "configs": [{"config_name": "afrikaans_ner_corpus", "data_files": [{"split": "train", "path": "afrikaans_ner_corpus/train-*"}], "default": true}]}
2024-01-09T11:51:47+00:00
[]
[ "af" ]
TAGS #task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Afrikaans #license-other #region-us
# Dataset Card for Afrikaans Ner Corpus ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: Afrikaans Ner Corpus Homepage - Repository: - Paper: - Leaderboard: - Point of Contact: Martin Puttkammer ### Dataset Summary The Afrikaans Ner Corpus is an Afrikaans dataset developed by The Centre for Text Technology (CTexT), North-West University, South Africa. The data is based on documents from the South African goverment domain and crawled from URL websites. It was created to support NER task for Afrikaans language. The dataset uses CoNLL shared task annotation standards. ### Supported Tasks and Leaderboards ### Languages The language supported is Afrikaans. ## Dataset Structure ### Data Instances A data point consists of sentences seperated by empty line and tab-seperated tokens and tags. {'id': '0', 'ner_tags': [0, 0, 0, 0, 0], 'tokens': ['Vertaling', 'van', 'die', 'inligting', 'in'] } ### Data Fields - 'id': id of the sample - 'tokens': the tokens of the example text - 'ner_tags': the NER tags of each token The NER tags correspond to this list: The NER tags have the same format as in the CoNLL shared task: a B denotes the first item of a phrase and an I any non-initial word. There are four types of phrases: person names (PER), organizations (ORG), locations (LOC) and miscellaneous names (MISC). (OUT) is used for tokens not considered part of any named entity. ### Data Splits The data was not split. ## Dataset Creation ### Curation Rationale The data was created to help introduce resources to new language - Afrikaans. ### Source Data #### Initial Data Collection and Normalization The data is based on South African government domain and was crawled from URL websites. #### Who are the source language producers? The data was produced by writers of South African government websites - URL ### Annotations #### Annotation process #### Who are the annotators? The data was annotated during the NCHLT text resource development project. ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators The annotated data sets were developed by the Centre for Text Technology (CTexT, North-West University, South Africa). See: more information ### Licensing Information The data is under the Creative Commons Attribution 2.5 South Africa License ### Contributions Thanks to @yvonnegitau for adding this dataset.
[ "# Dataset Card for Afrikaans Ner Corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Afrikaans Ner Corpus Homepage\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact: Martin Puttkammer", "### Dataset Summary\nThe Afrikaans Ner Corpus is an Afrikaans dataset developed by The Centre for Text Technology (CTexT), North-West University, South Africa. The data is based on documents from the South African goverment domain and crawled from URL websites. It was created to support NER task for Afrikaans language. The dataset uses CoNLL shared task annotation standards.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe language supported is Afrikaans.", "## Dataset Structure", "### Data Instances\n\nA data point consists of sentences seperated by empty line and tab-seperated tokens and tags. \n{'id': '0',\n 'ner_tags': [0, 0, 0, 0, 0],\n 'tokens': ['Vertaling', 'van', 'die', 'inligting', 'in']\n}", "### Data Fields\n\n- 'id': id of the sample\n- 'tokens': the tokens of the example text\n- 'ner_tags': the NER tags of each token\n\nThe NER tags correspond to this list:\n\nThe NER tags have the same format as in the CoNLL shared task: a B denotes the first item of a phrase and an I any non-initial word. There are four types of phrases: person names (PER), organizations (ORG), locations (LOC) and miscellaneous names (MISC). (OUT) is used for tokens not considered part of any named entity.", "### Data Splits\n\nThe data was not split.", "## Dataset Creation", "### Curation Rationale\n\nThe data was created to help introduce resources to new language - Afrikaans.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe data is based on South African government domain and was crawled from URL websites.", "#### Who are the source language producers?\n\nThe data was produced by writers of South African government websites - URL", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\nThe data was annotated during the NCHLT text resource development project.", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe annotated data sets were developed by the Centre for Text Technology (CTexT, North-West University, South Africa).\n\nSee: more information", "### Licensing Information\n\nThe data is under the Creative Commons Attribution 2.5 South Africa License", "### Contributions\n\nThanks to @yvonnegitau for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Afrikaans #license-other #region-us \n", "# Dataset Card for Afrikaans Ner Corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Afrikaans Ner Corpus Homepage\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact: Martin Puttkammer", "### Dataset Summary\nThe Afrikaans Ner Corpus is an Afrikaans dataset developed by The Centre for Text Technology (CTexT), North-West University, South Africa. The data is based on documents from the South African goverment domain and crawled from URL websites. It was created to support NER task for Afrikaans language. The dataset uses CoNLL shared task annotation standards.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe language supported is Afrikaans.", "## Dataset Structure", "### Data Instances\n\nA data point consists of sentences seperated by empty line and tab-seperated tokens and tags. \n{'id': '0',\n 'ner_tags': [0, 0, 0, 0, 0],\n 'tokens': ['Vertaling', 'van', 'die', 'inligting', 'in']\n}", "### Data Fields\n\n- 'id': id of the sample\n- 'tokens': the tokens of the example text\n- 'ner_tags': the NER tags of each token\n\nThe NER tags correspond to this list:\n\nThe NER tags have the same format as in the CoNLL shared task: a B denotes the first item of a phrase and an I any non-initial word. There are four types of phrases: person names (PER), organizations (ORG), locations (LOC) and miscellaneous names (MISC). (OUT) is used for tokens not considered part of any named entity.", "### Data Splits\n\nThe data was not split.", "## Dataset Creation", "### Curation Rationale\n\nThe data was created to help introduce resources to new language - Afrikaans.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe data is based on South African government domain and was crawled from URL websites.", "#### Who are the source language producers?\n\nThe data was produced by writers of South African government websites - URL", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\nThe data was annotated during the NCHLT text resource development project.", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe annotated data sets were developed by the Centre for Text Technology (CTexT, North-West University, South Africa).\n\nSee: more information", "### Licensing Information\n\nThe data is under the Creative Commons Attribution 2.5 South Africa License", "### Contributions\n\nThanks to @yvonnegitau for adding this dataset." ]
68a83b6cd4730be5e0ecbdbee941eef8f13aa867
# Dataset Card for "ag_news" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [http://groups.di.unipi.it/~gulli/AG_corpus_of_news_articles.html](http://groups.di.unipi.it/~gulli/AG_corpus_of_news_articles.html) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 31.33 MB - **Size of the generated dataset:** 31.70 MB - **Total amount of disk used:** 63.02 MB ### Dataset Summary AG is a collection of more than 1 million news articles. News articles have been gathered from more than 2000 news sources by ComeToMyHead in more than 1 year of activity. ComeToMyHead is an academic news search engine which has been running since July, 2004. The dataset is provided by the academic comunity for research purposes in data mining (clustering, classification, etc), information retrieval (ranking, search, etc), xml, data compression, data streaming, and any other non-commercial activity. For more information, please refer to the link http://www.di.unipi.it/~gulli/AG_corpus_of_news_articles.html . The AG's news topic classification dataset is constructed by Xiang Zhang (xiang.zhang@nyu.edu) from the dataset above. It is used as a text classification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann LeCun. Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015). ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 31.33 MB - **Size of the generated dataset:** 31.70 MB - **Total amount of disk used:** 63.02 MB An example of 'train' looks as follows. ``` { "label": 3, "text": "New iPad released Just like every other September, this one is no different. Apple is planning to release a bigger, heavier, fatter iPad that..." } ``` ### Data Fields The data fields are the same among all splits. #### default - `text`: a `string` feature. - `label`: a classification label, with possible values including `World` (0), `Sports` (1), `Business` (2), `Sci/Tech` (3). ### Data Splits | name |train |test| |-------|-----:|---:| |default|120000|7600| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @inproceedings{Zhang2015CharacterlevelCN, title={Character-level Convolutional Networks for Text Classification}, author={Xiang Zhang and Junbo Jake Zhao and Yann LeCun}, booktitle={NIPS}, year={2015} } ``` ### Contributions Thanks to [@jxmorris12](https://github.com/jxmorris12), [@thomwolf](https://github.com/thomwolf), [@lhoestq](https://github.com/lhoestq), [@lewtun](https://github.com/lewtun) for adding this dataset.
ag_news
[ "task_categories:text-classification", "task_ids:topic-classification", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["topic-classification"], "paperswithcode_id": "ag-news", "pretty_name": "AG\u2019s News Corpus", "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "World", "1": "Sports", "2": "Business", "3": "Sci/Tech"}}}}], "splits": [{"name": "train", "num_bytes": 29817351, "num_examples": 120000}, {"name": "test", "num_bytes": 1879478, "num_examples": 7600}], "download_size": 31327765, "dataset_size": 31696829}, "train-eval-index": [{"config": "default", "task": "text-classification", "task_id": "multi_class_classification", "splits": {"train_split": "train", "eval_split": "test"}, "col_mapping": {"text": "text", "label": "target"}, "metrics": [{"type": "accuracy", "name": "Accuracy"}, {"type": "f1", "name": "F1 macro", "args": {"average": "macro"}}, {"type": "f1", "name": "F1 micro", "args": {"average": "micro"}}, {"type": "f1", "name": "F1 weighted", "args": {"average": "weighted"}}, {"type": "precision", "name": "Precision macro", "args": {"average": "macro"}}, {"type": "precision", "name": "Precision micro", "args": {"average": "micro"}}, {"type": "precision", "name": "Precision weighted", "args": {"average": "weighted"}}, {"type": "recall", "name": "Recall macro", "args": {"average": "macro"}}, {"type": "recall", "name": "Recall micro", "args": {"average": "micro"}}, {"type": "recall", "name": "Recall weighted", "args": {"average": "weighted"}}]}]}
2024-01-18T10:52:09+00:00
[]
[ "en" ]
TAGS #task_categories-text-classification #task_ids-topic-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #region-us
Dataset Card for "ag\_news" =========================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: * Size of downloaded dataset files: 31.33 MB * Size of the generated dataset: 31.70 MB * Total amount of disk used: 63.02 MB ### Dataset Summary AG is a collection of more than 1 million news articles. News articles have been gathered from more than 2000 news sources by ComeToMyHead in more than 1 year of activity. ComeToMyHead is an academic news search engine which has been running since July, 2004. The dataset is provided by the academic comunity for research purposes in data mining (clustering, classification, etc), information retrieval (ranking, search, etc), xml, data compression, data streaming, and any other non-commercial activity. For more information, please refer to the link URL . The AG's news topic classification dataset is constructed by Xiang Zhang (URL@URL) from the dataset above. It is used as a text classification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann LeCun. Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015). ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### default * Size of downloaded dataset files: 31.33 MB * Size of the generated dataset: 31.70 MB * Total amount of disk used: 63.02 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### default * 'text': a 'string' feature. * 'label': a classification label, with possible values including 'World' (0), 'Sports' (1), 'Business' (2), 'Sci/Tech' (3). ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @jxmorris12, @thomwolf, @lhoestq, @lewtun for adding this dataset.
[ "### Dataset Summary\n\n\nAG is a collection of more than 1 million news articles. News articles have been\ngathered from more than 2000 news sources by ComeToMyHead in more than 1 year of\nactivity. ComeToMyHead is an academic news search engine which has been running\nsince July, 2004. The dataset is provided by the academic comunity for research\npurposes in data mining (clustering, classification, etc), information retrieval\n(ranking, search, etc), xml, data compression, data streaming, and any other\nnon-commercial activity. For more information, please refer to the link\nURL .\n\n\nThe AG's news topic classification dataset is constructed by Xiang Zhang\n(URL@URL) from the dataset above. It is used as a text\nclassification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann\nLeCun. Character-level Convolutional Networks for Text Classification. Advances\nin Neural Information Processing Systems 28 (NIPS 2015).", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 31.33 MB\n* Size of the generated dataset: 31.70 MB\n* Total amount of disk used: 63.02 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'text': a 'string' feature.\n* 'label': a classification label, with possible values including 'World' (0), 'Sports' (1), 'Business' (2), 'Sci/Tech' (3).", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @jxmorris12, @thomwolf, @lhoestq, @lewtun for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-topic-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #region-us \n", "### Dataset Summary\n\n\nAG is a collection of more than 1 million news articles. News articles have been\ngathered from more than 2000 news sources by ComeToMyHead in more than 1 year of\nactivity. ComeToMyHead is an academic news search engine which has been running\nsince July, 2004. The dataset is provided by the academic comunity for research\npurposes in data mining (clustering, classification, etc), information retrieval\n(ranking, search, etc), xml, data compression, data streaming, and any other\nnon-commercial activity. For more information, please refer to the link\nURL .\n\n\nThe AG's news topic classification dataset is constructed by Xiang Zhang\n(URL@URL) from the dataset above. It is used as a text\nclassification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann\nLeCun. Character-level Convolutional Networks for Text Classification. Advances\nin Neural Information Processing Systems 28 (NIPS 2015).", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 31.33 MB\n* Size of the generated dataset: 31.70 MB\n* Total amount of disk used: 63.02 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'text': a 'string' feature.\n* 'label': a classification label, with possible values including 'World' (0), 'Sports' (1), 'Business' (2), 'Sci/Tech' (3).", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @jxmorris12, @thomwolf, @lhoestq, @lewtun for adding this dataset." ]
210d026faf9955653af8916fad021475a3f00453
# Dataset Card for "ai2_arc" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://allenai.org/data/arc](https://allenai.org/data/arc) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge](https://arxiv.org/abs/1803.05457) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 1361.68 MB - **Size of the generated dataset:** 2.28 MB - **Total amount of disk used:** 1363.96 MB ### Dataset Summary A new dataset of 7,787 genuine grade-school level, multiple-choice science questions, assembled to encourage research in advanced question-answering. The dataset is partitioned into a Challenge Set and an Easy Set, where the former contains only questions answered incorrectly by both a retrieval-based algorithm and a word co-occurrence algorithm. We are also including a corpus of over 14 million science sentences relevant to the task, and an implementation of three neural baseline models for this dataset. We pose ARC as a challenge to the community. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### ARC-Challenge - **Size of downloaded dataset files:** 680.84 MB - **Size of the generated dataset:** 0.83 MB - **Total amount of disk used:** 681.67 MB An example of 'train' looks as follows. ``` { "answerKey": "B", "choices": { "label": ["A", "B", "C", "D"], "text": ["Shady areas increased.", "Food sources increased.", "Oxygen levels increased.", "Available water increased."] }, "id": "Mercury_SC_405487", "question": "One year, the oak trees in a park began producing more acorns than usual. The next year, the population of chipmunks in the park also increased. Which best explains why there were more chipmunks the next year?" } ``` #### ARC-Easy - **Size of downloaded dataset files:** 680.84 MB - **Size of the generated dataset:** 1.45 MB - **Total amount of disk used:** 682.29 MB An example of 'train' looks as follows. ``` { "answerKey": "B", "choices": { "label": ["A", "B", "C", "D"], "text": ["Shady areas increased.", "Food sources increased.", "Oxygen levels increased.", "Available water increased."] }, "id": "Mercury_SC_405487", "question": "One year, the oak trees in a park began producing more acorns than usual. The next year, the population of chipmunks in the park also increased. Which best explains why there were more chipmunks the next year?" } ``` ### Data Fields The data fields are the same among all splits. #### ARC-Challenge - `id`: a `string` feature. - `question`: a `string` feature. - `choices`: a dictionary feature containing: - `text`: a `string` feature. - `label`: a `string` feature. - `answerKey`: a `string` feature. #### ARC-Easy - `id`: a `string` feature. - `question`: a `string` feature. - `choices`: a dictionary feature containing: - `text`: a `string` feature. - `label`: a `string` feature. - `answerKey`: a `string` feature. ### Data Splits | name |train|validation|test| |-------------|----:|---------:|---:| |ARC-Challenge| 1119| 299|1172| |ARC-Easy | 2251| 570|2376| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @article{allenai:arc, author = {Peter Clark and Isaac Cowhey and Oren Etzioni and Tushar Khot and Ashish Sabharwal and Carissa Schoenick and Oyvind Tafjord}, title = {Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge}, journal = {arXiv:1803.05457v1}, year = {2018}, } ``` ### Contributions Thanks to [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf) for adding this dataset.
allenai/ai2_arc
[ "task_categories:question-answering", "task_ids:open-domain-qa", "task_ids:multiple-choice-qa", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "arxiv:1803.05457", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["question-answering"], "task_ids": ["open-domain-qa", "multiple-choice-qa"], "pretty_name": "Ai2Arc", "language_bcp47": ["en-US"], "dataset_info": [{"config_name": "ARC-Challenge", "features": [{"name": "id", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "choices", "sequence": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": "string"}]}, {"name": "answerKey", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 349760, "num_examples": 1119}, {"name": "test", "num_bytes": 375511, "num_examples": 1172}, {"name": "validation", "num_bytes": 96660, "num_examples": 299}], "download_size": 449460, "dataset_size": 821931}, {"config_name": "ARC-Easy", "features": [{"name": "id", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "choices", "sequence": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": "string"}]}, {"name": "answerKey", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 619000, "num_examples": 2251}, {"name": "test", "num_bytes": 657514, "num_examples": 2376}, {"name": "validation", "num_bytes": 157394, "num_examples": 570}], "download_size": 762935, "dataset_size": 1433908}], "configs": [{"config_name": "ARC-Challenge", "data_files": [{"split": "train", "path": "ARC-Challenge/train-*"}, {"split": "test", "path": "ARC-Challenge/test-*"}, {"split": "validation", "path": "ARC-Challenge/validation-*"}]}, {"config_name": "ARC-Easy", "data_files": [{"split": "train", "path": "ARC-Easy/train-*"}, {"split": "test", "path": "ARC-Easy/test-*"}, {"split": "validation", "path": "ARC-Easy/validation-*"}]}]}
2023-12-21T15:09:48+00:00
[ "1803.05457" ]
[ "en" ]
TAGS #task_categories-question-answering #task_ids-open-domain-qa #task_ids-multiple-choice-qa #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-1803.05457 #region-us
Dataset Card for "ai2\_arc" =========================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge * Point of Contact: * Size of downloaded dataset files: 1361.68 MB * Size of the generated dataset: 2.28 MB * Total amount of disk used: 1363.96 MB ### Dataset Summary A new dataset of 7,787 genuine grade-school level, multiple-choice science questions, assembled to encourage research in advanced question-answering. The dataset is partitioned into a Challenge Set and an Easy Set, where the former contains only questions answered incorrectly by both a retrieval-based algorithm and a word co-occurrence algorithm. We are also including a corpus of over 14 million science sentences relevant to the task, and an implementation of three neural baseline models for this dataset. We pose ARC as a challenge to the community. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### ARC-Challenge * Size of downloaded dataset files: 680.84 MB * Size of the generated dataset: 0.83 MB * Total amount of disk used: 681.67 MB An example of 'train' looks as follows. #### ARC-Easy * Size of downloaded dataset files: 680.84 MB * Size of the generated dataset: 1.45 MB * Total amount of disk used: 682.29 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### ARC-Challenge * 'id': a 'string' feature. * 'question': a 'string' feature. * 'choices': a dictionary feature containing: + 'text': a 'string' feature. + 'label': a 'string' feature. * 'answerKey': a 'string' feature. #### ARC-Easy * 'id': a 'string' feature. * 'question': a 'string' feature. * 'choices': a dictionary feature containing: + 'text': a 'string' feature. + 'label': a 'string' feature. * 'answerKey': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @lewtun, @patrickvonplaten, @thomwolf for adding this dataset.
[ "### Dataset Summary\n\n\nA new dataset of 7,787 genuine grade-school level, multiple-choice science questions, assembled to encourage research in\nadvanced question-answering. The dataset is partitioned into a Challenge Set and an Easy Set, where the former contains\nonly questions answered incorrectly by both a retrieval-based algorithm and a word co-occurrence algorithm. We are also\nincluding a corpus of over 14 million science sentences relevant to the task, and an implementation of three neural baseline models for this dataset. We pose ARC as a challenge to the community.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### ARC-Challenge\n\n\n* Size of downloaded dataset files: 680.84 MB\n* Size of the generated dataset: 0.83 MB\n* Total amount of disk used: 681.67 MB\n\n\nAn example of 'train' looks as follows.", "#### ARC-Easy\n\n\n* Size of downloaded dataset files: 680.84 MB\n* Size of the generated dataset: 1.45 MB\n* Total amount of disk used: 682.29 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### ARC-Challenge\n\n\n* 'id': a 'string' feature.\n* 'question': a 'string' feature.\n* 'choices': a dictionary feature containing:\n\t+ 'text': a 'string' feature.\n\t+ 'label': a 'string' feature.\n* 'answerKey': a 'string' feature.", "#### ARC-Easy\n\n\n* 'id': a 'string' feature.\n* 'question': a 'string' feature.\n* 'choices': a dictionary feature containing:\n\t+ 'text': a 'string' feature.\n\t+ 'label': a 'string' feature.\n* 'answerKey': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @lewtun, @patrickvonplaten, @thomwolf for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-open-domain-qa #task_ids-multiple-choice-qa #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-1803.05457 #region-us \n", "### Dataset Summary\n\n\nA new dataset of 7,787 genuine grade-school level, multiple-choice science questions, assembled to encourage research in\nadvanced question-answering. The dataset is partitioned into a Challenge Set and an Easy Set, where the former contains\nonly questions answered incorrectly by both a retrieval-based algorithm and a word co-occurrence algorithm. We are also\nincluding a corpus of over 14 million science sentences relevant to the task, and an implementation of three neural baseline models for this dataset. We pose ARC as a challenge to the community.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### ARC-Challenge\n\n\n* Size of downloaded dataset files: 680.84 MB\n* Size of the generated dataset: 0.83 MB\n* Total amount of disk used: 681.67 MB\n\n\nAn example of 'train' looks as follows.", "#### ARC-Easy\n\n\n* Size of downloaded dataset files: 680.84 MB\n* Size of the generated dataset: 1.45 MB\n* Total amount of disk used: 682.29 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### ARC-Challenge\n\n\n* 'id': a 'string' feature.\n* 'question': a 'string' feature.\n* 'choices': a dictionary feature containing:\n\t+ 'text': a 'string' feature.\n\t+ 'label': a 'string' feature.\n* 'answerKey': a 'string' feature.", "#### ARC-Easy\n\n\n* 'id': a 'string' feature.\n* 'question': a 'string' feature.\n* 'choices': a dictionary feature containing:\n\t+ 'text': a 'string' feature.\n\t+ 'label': a 'string' feature.\n* 'answerKey': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @lewtun, @patrickvonplaten, @thomwolf for adding this dataset." ]
69a8c7b33b9ae3281d93bdc34e85735b2ad4e662
# Dataset Card for air_dialogue ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://worksheets.codalab.org/worksheets/0xa79833f4b3c24f4188cee7131b120a59 - **Repository:** https://github.com/google/airdialogue - **Paper:** https://www.aclweb.org/anthology/D18-1419/ - **Leaderboard:** https://worksheets.codalab.org/worksheets/0xa79833f4b3c24f4188cee7131b120a59 - **Point of Contact:** [AirDialogue-Google](mailto:airdialogue@gmail.com) [Aakash Gupta](mailto:aakashg80@gmail.com) ### Dataset Summary AirDialogue, is a large dataset that contains 402,038 goal-oriented conversations. To collect this dataset, we create a contextgenerator which provides travel and flight restrictions. Then the human annotators are asked to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. ### Supported Tasks and Leaderboards We use perplexity and BLEU score to evaluate the quality of the language generated by the model. We also compare the dialogue state generated by the model s and the ground truth state s0. Two categories of the metrics are used: exact match scores and scaled scores The inference competition & leaderboard can be found here: https://worksheets.codalab.org/worksheets/0xa79833f4b3c24f4188cee7131b120a59 ### Languages The text in the dataset is in English. The BCP 47 code is `en` ## Dataset Structure ### Data Instances The data is provided in two set of files. The first one has the dialogues (`air_dialogue_data`) and the knowledge-base (`air_dialogue_kb`) BuilderConfig: `air_dialogue_data` ``` {"action": {"status": "book", "name": "Emily Edwards", "flight": [1027]}, "intent": {"return_month": "June", "return_day": "14", "max_price": 200, "departure_airport": "DFW", "return_time": "afternoon", "max_connections": 1, "departure_day": "12", "goal": "book", "departure_month": "June", "name": "Emily Edwards", "return_airport": "IAD"}, "timestamps": [1519233239, 1519233244, 1519233249, 1519233252, 1519233333, 1519233374, 1519233392, 1519233416, 1519233443, 1519233448, 1519233464, 1519233513, 1519233525, 1519233540, 1519233626, 1519233628, 1519233638], "dialogue": ["customer: Hello.", "agent: Hello.", "customer: My name is Emily Edwards.", "agent: How may I help you out?", "customer: I need some help in my flight ticket reservation to attend a convocation meeting, can you please help me?", "agent: Sure, I will help you out. May I know your travelling dates please?", "customer: Thank you and my dates are 06/12 and back on 06/14.", "agent: Can I know your airport codes?", "customer: The airport codes are from DFW to IAD.", "agent: Ok, please wait a moment.", "customer: Sure.", "agent: There is a flight with connection 1 and price 200, can I proceed with this flight?", "customer: Yes, do proceed with booking.", "agent: Ok, your ticket has been booked.", "customer: Thank you for your assistance in my flight ticket reservation.", "agent: Thank you for choosing us.", "customer: You are welcome."], "expected_action": {"status": "book", "name": "Emily Edwards", "flight": [1027]}, "correct_sample": true} ``` BuilderConfig: `air_dialogue_kb` ``` {"kb": [{"return_airport": "DTW", "airline": "Spirit", "departure_day": "12", "departure_airport": "IAD", "flight_number": 1000, "departure_month": "June", "departure_time_num": 17, "class": "economy", "return_time_num": 2, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 200}, {"return_airport": "DTW", "airline": "Frontier", "departure_day": "12", "departure_airport": "IAD", "flight_number": 1001, "departure_month": "June", "departure_time_num": 0, "class": "business", "return_time_num": 15, "return_month": "June", "return_day": "13", "num_connections": 0, "price": 500}, {"return_airport": "DTW", "airline": "JetBlue", "departure_day": "12", "departure_airport": "IAD", "flight_number": 1002, "departure_month": "June", "departure_time_num": 0, "class": "business", "return_time_num": 13, "return_month": "June", "return_day": "13", "num_connections": 1, "price": 600}, {"return_airport": "IAD", "airline": "Hawaiian", "departure_day": "12", "departure_airport": "DTW", "flight_number": 1003, "departure_month": "June", "departure_time_num": 6, "class": "economy", "return_time_num": 5, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 200}, {"return_airport": "DFW", "airline": "AA", "departure_day": "12", "departure_airport": "DTW", "flight_number": 1004, "departure_month": "June", "departure_time_num": 9, "class": "economy", "return_time_num": 11, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 100}, {"return_airport": "IAD", "airline": "AA", "departure_day": "12", "departure_airport": "DFW", "flight_number": 1005, "departure_month": "June", "departure_time_num": 3, "class": "economy", "return_time_num": 17, "return_month": "June", "return_day": "13", "num_connections": 1, "price": 100}, {"return_airport": "DTW", "airline": "Frontier", "departure_day": "12", "departure_airport": "IAD", "flight_number": 1006, "departure_month": "June", "departure_time_num": 10, "class": "economy", "return_time_num": 10, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 100}, {"return_airport": "IAD", "airline": "UA", "departure_day": "12", "departure_airport": "DFW", "flight_number": 1007, "departure_month": "June", "departure_time_num": 14, "class": "economy", "return_time_num": 20, "return_month": "June", "return_day": "13", "num_connections": 1, "price": 100}, {"return_airport": "DFW", "airline": "AA", "departure_day": "13", "departure_airport": "DTW", "flight_number": 1008, "departure_month": "June", "departure_time_num": 6, "class": "economy", "return_time_num": 8, "return_month": "June", "return_day": "14", "num_connections": 2, "price": 400}, {"return_airport": "DFW", "airline": "Delta", "departure_day": "12", "departure_airport": "IAD", "flight_number": 1009, "departure_month": "June", "departure_time_num": 18, "class": "economy", "return_time_num": 6, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 200}, {"return_airport": "DFW", "airline": "Frontier", "departure_day": "13", "departure_airport": "DTW", "flight_number": 1010, "departure_month": "June", "departure_time_num": 4, "class": "economy", "return_time_num": 2, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 100}, {"return_airport": "DFW", "airline": "Southwest", "departure_day": "12", "departure_airport": "DTW", "flight_number": 1011, "departure_month": "June", "departure_time_num": 17, "class": "economy", "return_time_num": 22, "return_month": "June", "return_day": "13", "num_connections": 0, "price": 100}, {"return_airport": "DTW", "airline": "JetBlue", "departure_day": "11", "departure_airport": "DFW", "flight_number": 1012, "departure_month": "June", "departure_time_num": 13, "class": "economy", "return_time_num": 22, "return_month": "June", "return_day": "13", "num_connections": 1, "price": 100}, {"return_airport": "DTW", "airline": "Southwest", "departure_day": "12", "departure_airport": "IAD", "flight_number": 1013, "departure_month": "June", "departure_time_num": 16, "class": "economy", "return_time_num": 13, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 200}, {"return_airport": "DTW", "airline": "Delta", "departure_day": "12", "departure_airport": "IAD", "flight_number": 1014, "departure_month": "June", "departure_time_num": 0, "class": "economy", "return_time_num": 8, "return_month": "June", "return_day": "15", "num_connections": 1, "price": 100}, {"return_airport": "DTW", "airline": "Southwest", "departure_day": "12", "departure_airport": "DFW", "flight_number": 1015, "departure_month": "June", "departure_time_num": 17, "class": "economy", "return_time_num": 1, "return_month": "June", "return_day": "15", "num_connections": 1, "price": 300}, {"return_airport": "DTW", "airline": "UA", "departure_day": "11", "departure_airport": "DFW", "flight_number": 1016, "departure_month": "June", "departure_time_num": 10, "class": "economy", "return_time_num": 4, "return_month": "June", "return_day": "14", "num_connections": 0, "price": 200}, {"return_airport": "DFW", "airline": "AA", "departure_day": "12", "departure_airport": "DTW", "flight_number": 1017, "departure_month": "June", "departure_time_num": 14, "class": "economy", "return_time_num": 23, "return_month": "June", "return_day": "14", "num_connections": 2, "price": 400}, {"return_airport": "DTW", "airline": "JetBlue", "departure_day": "12", "departure_airport": "DFW", "flight_number": 1018, "departure_month": "June", "departure_time_num": 3, "class": "economy", "return_time_num": 1, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 100}, {"return_airport": "DFW", "airline": "Hawaiian", "departure_day": "12", "departure_airport": "IAD", "flight_number": 1019, "departure_month": "June", "departure_time_num": 7, "class": "economy", "return_time_num": 18, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 200}, {"return_airport": "DFW", "airline": "Delta", "departure_day": "12", "departure_airport": "IAD", "flight_number": 1020, "departure_month": "June", "departure_time_num": 6, "class": "economy", "return_time_num": 18, "return_month": "June", "return_day": "14", "num_connections": 2, "price": 200}, {"return_airport": "IAD", "airline": "Delta", "departure_day": "12", "departure_airport": "DFW", "flight_number": 1021, "departure_month": "June", "departure_time_num": 11, "class": "business", "return_time_num": 8, "return_month": "June", "return_day": "14", "num_connections": 0, "price": 1000}, {"return_airport": "IAD", "airline": "JetBlue", "departure_day": "12", "departure_airport": "DTW", "flight_number": 1022, "departure_month": "June", "departure_time_num": 4, "class": "economy", "return_time_num": 14, "return_month": "June", "return_day": "13", "num_connections": 0, "price": 200}, {"return_airport": "IAD", "airline": "Frontier", "departure_day": "12", "departure_airport": "DTW", "flight_number": 1023, "departure_month": "June", "departure_time_num": 19, "class": "economy", "return_time_num": 23, "return_month": "June", "return_day": "13", "num_connections": 1, "price": 200}, {"return_airport": "DFW", "airline": "UA", "departure_day": "12", "departure_airport": "DTW", "flight_number": 1024, "departure_month": "June", "departure_time_num": 11, "class": "economy", "return_time_num": 19, "return_month": "June", "return_day": "15", "num_connections": 1, "price": 200}, {"return_airport": "DTW", "airline": "Hawaiian", "departure_day": "11", "departure_airport": "IAD", "flight_number": 1025, "departure_month": "June", "departure_time_num": 6, "class": "economy", "return_time_num": 10, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 100}, {"return_airport": "DTW", "airline": "UA", "departure_day": "12", "departure_airport": "DFW", "flight_number": 1026, "departure_month": "June", "departure_time_num": 0, "class": "economy", "return_time_num": 18, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 300}, {"return_airport": "IAD", "airline": "Delta", "departure_day": "12", "departure_airport": "DFW", "flight_number": 1027, "departure_month": "June", "departure_time_num": 17, "class": "economy", "return_time_num": 15, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 200}, {"return_airport": "IAD", "airline": "Southwest", "departure_day": "12", "departure_airport": "DTW", "flight_number": 1028, "departure_month": "June", "departure_time_num": 23, "class": "economy", "return_time_num": 13, "return_month": "June", "return_day": "14", "num_connections": 1, "price": 100}, {"return_airport": "DFW", "airline": "Spirit", "departure_day": "11", "departure_airport": "DTW", "flight_number": 1029, "departure_month": "June", "departure_time_num": 22, "class": "business", "return_time_num": 4, "return_month": "June", "return_day": "14", "num_connections": 0, "price": 800}], "reservation": 0} ``` ### Data Fields BuilderConfig: `air_dialogue_data`: Provides for customer context, dialogue states and environment key name | Description | |---|---| |'search_action' | search action performed by customer | |'action' | Action taken by the agent | |'intent' | Intents from the conversation | |'timestamps' | Timestamp for each of the dialogues | |'dialogue' | Dialogue recorded between agent & customer | |'expected_action' | Expected action from agent (human-annotated)| |'correct_sample' | whether action performed by agent was same as expected_action | BuilderConfig: `air_dialogue_kb`: Provides for the Agent Context _ca_ = (_db_, _r_ ) key name | Description | |---|---| |'kb' | Available flights in the database | |'reservation' | whether customer has an existing reservation| ### Data Splits Data is split into Train/Dev & Test in the ration of 80%, 10% and 10% ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process To collect this dataset, we create a contextgenerator which provides travel and flight restrictions. We then ask human annotators to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. Key to our environment is the ease of evaluating the success of the dialogue, which is achieved by using ground-truth states (e.g., the flight being booked) generated by the restrictions. Any dialogue agent that does not generate the correct states is considered to fail. #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information No personal and sensitive information is stored ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [AirDialogue team](mailto:airdialogue@gmail.com) For issues regarding HuggingFace Dataset Hub implementation [Aakash Gupta](mailto:aakashg80@gmail.com) ### Licensing Information cc-by-nc-4.0 ### Citation Information @inproceedings{wei-etal-2018-airdialogue, title = "{A}ir{D}ialogue: An Environment for Goal-Oriented Dialogue Research", author = "Wei, Wei and Le, Quoc and Dai, Andrew and Li, Jia", booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", month = oct # "-" # nov, year = "2018", address = "Brussels, Belgium", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/D18-1419", doi = "10.18653/v1/D18-1419", pages = "3844--3854", abstract = "Recent progress in dialogue generation has inspired a number of studies on dialogue systems that are capable of accomplishing tasks through natural language interactions. A promising direction among these studies is the use of reinforcement learning techniques, such as self-play, for training dialogue agents. However, current datasets are limited in size, and the environment for training agents and evaluating progress is relatively unsophisticated. We present AirDialogue, a large dataset that contains 301,427 goal-oriented conversations. To collect this dataset, we create a context-generator which provides travel and flight restrictions. We then ask human annotators to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. Key to our environment is the ease of evaluating the success of the dialogue, which is achieved by using ground-truth states (e.g., the flight being booked) generated by the restrictions. Any dialogue agent that does not generate the correct states is considered to fail. Our experimental results indicate that state-of-the-art dialogue models can only achieve a score of 0.17 while humans can reach a score of 0.91, which suggests significant opportunities for future improvement.", } ### Contributions Thanks to [@skyprince999](https://github.com/skyprince999) for adding this dataset.
air_dialogue
[ "task_categories:conversational", "task_categories:text-generation", "task_categories:fill-mask", "task_ids:dialogue-generation", "task_ids:dialogue-modeling", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:crowdsourced", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:cc-by-nc-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["machine-generated"], "language": ["en"], "license": ["cc-by-nc-4.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["conversational", "text-generation", "fill-mask"], "task_ids": ["dialogue-generation", "dialogue-modeling", "language-modeling", "masked-language-modeling"], "pretty_name": "AirDialogue", "dataset_info": [{"config_name": "air_dialogue_data", "features": [{"name": "action", "struct": [{"name": "status", "dtype": "string"}, {"name": "name", "dtype": "string"}, {"name": "flight", "sequence": "int32"}]}, {"name": "intent", "struct": [{"name": "return_month", "dtype": "string"}, {"name": "return_day", "dtype": "string"}, {"name": "max_price", "dtype": "int32"}, {"name": "departure_airport", "dtype": "string"}, {"name": "max_connections", "dtype": "int32"}, {"name": "departure_day", "dtype": "string"}, {"name": "goal", "dtype": "string"}, {"name": "departure_month", "dtype": "string"}, {"name": "name", "dtype": "string"}, {"name": "return_airport", "dtype": "string"}]}, {"name": "timestamps", "sequence": "int64"}, {"name": "dialogue", "sequence": "string"}, {"name": "expected_action", "struct": [{"name": "status", "dtype": "string"}, {"name": "name", "dtype": "string"}, {"name": "flight", "sequence": "int32"}]}, {"name": "search_info", "list": [{"name": "button_name", "dtype": "string"}, {"name": "field_name", "dtype": "string"}, {"name": "field_value", "dtype": "string"}, {"name": "timestmamp", "dtype": "int64"}]}, {"name": "correct_sample", "dtype": "bool_"}], "splits": [{"name": "train", "num_bytes": 353721137, "num_examples": 321459}, {"name": "validation", "num_bytes": 44442238, "num_examples": 40363}], "download_size": 272898923, "dataset_size": 398163375}, {"config_name": "air_dialogue_kb", "features": [{"name": "kb", "list": [{"name": "airline", "dtype": "string"}, {"name": "class", "dtype": "string"}, {"name": "departure_airport", "dtype": "string"}, {"name": "departure_day", "dtype": "string"}, {"name": "departure_month", "dtype": "string"}, {"name": "departure_time_num", "dtype": "int32"}, {"name": "flight_number", "dtype": "int32"}, {"name": "num_connections", "dtype": "int32"}, {"name": "price", "dtype": "int32"}, {"name": "return_airport", "dtype": "string"}, {"name": "return_day", "dtype": "string"}, {"name": "return_month", "dtype": "string"}, {"name": "return_time_num", "dtype": "int32"}]}, {"name": "reservation", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 782592158, "num_examples": 321459}, {"name": "validation", "num_bytes": 98269789, "num_examples": 40363}], "download_size": 272898923, "dataset_size": 880861947}]}
2024-01-18T10:59:29+00:00
[]
[ "en" ]
TAGS #task_categories-conversational #task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-generation #task_ids-dialogue-modeling #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-crowdsourced #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-nc-4.0 #region-us
Dataset Card for air\_dialogue ============================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: URL * Paper: URL * Leaderboard: URL * Point of Contact: AirDialogue-Google Aakash Gupta ### Dataset Summary AirDialogue, is a large dataset that contains 402,038 goal-oriented conversations. To collect this dataset, we create a contextgenerator which provides travel and flight restrictions. Then the human annotators are asked to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. ### Supported Tasks and Leaderboards We use perplexity and BLEU score to evaluate the quality of the language generated by the model. We also compare the dialogue state generated by the model s and the ground truth state s0. Two categories of the metrics are used: exact match scores and scaled scores The inference competition & leaderboard can be found here: URL ### Languages The text in the dataset is in English. The BCP 47 code is 'en' Dataset Structure ----------------- ### Data Instances The data is provided in two set of files. The first one has the dialogues ('air\_dialogue\_data') and the knowledge-base ('air\_dialogue\_kb') BuilderConfig: 'air\_dialogue\_data' BuilderConfig: 'air\_dialogue\_kb' ### Data Fields BuilderConfig: 'air\_dialogue\_data': Provides for customer context, dialogue states and environment BuilderConfig: 'air\_dialogue\_kb': Provides for the Agent Context *ca* = (*db*, *r* ) ### Data Splits Data is split into Train/Dev & Test in the ration of 80%, 10% and 10% Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process To collect this dataset, we create a contextgenerator which provides travel and flight restrictions. We then ask human annotators to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. Key to our environment is the ease of evaluating the success of the dialogue, which is achieved by using ground-truth states (e.g., the flight being booked) generated by the restrictions. Any dialogue agent that does not generate the correct states is considered to fail. #### Who are the annotators? ### Personal and Sensitive Information No personal and sensitive information is stored Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators AirDialogue team For issues regarding HuggingFace Dataset Hub implementation Aakash Gupta ### Licensing Information cc-by-nc-4.0 @inproceedings{wei-etal-2018-airdialogue, title = "{A}ir{D}ialogue: An Environment for Goal-Oriented Dialogue Research", author = "Wei, Wei and Le, Quoc and Dai, Andrew and Li, Jia", booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", month = oct # "-" # nov, year = "2018", address = "Brussels, Belgium", publisher = "Association for Computational Linguistics", url = "URL doi = "10.18653/v1/D18-1419", pages = "3844--3854", abstract = "Recent progress in dialogue generation has inspired a number of studies on dialogue systems that are capable of accomplishing tasks through natural language interactions. A promising direction among these studies is the use of reinforcement learning techniques, such as self-play, for training dialogue agents. However, current datasets are limited in size, and the environment for training agents and evaluating progress is relatively unsophisticated. We present AirDialogue, a large dataset that contains 301,427 goal-oriented conversations. To collect this dataset, we create a context-generator which provides travel and flight restrictions. We then ask human annotators to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. Key to our environment is the ease of evaluating the success of the dialogue, which is achieved by using ground-truth states (e.g., the flight being booked) generated by the restrictions. Any dialogue agent that does not generate the correct states is considered to fail. Our experimental results indicate that state-of-the-art dialogue models can only achieve a score of 0.17 while humans can reach a score of 0.91, which suggests significant opportunities for future improvement.", } ### Contributions Thanks to @skyprince999 for adding this dataset.
[ "### Dataset Summary\n\n\nAirDialogue, is a large dataset that contains 402,038 goal-oriented conversations. To collect this dataset, we create a contextgenerator which provides travel and flight restrictions. Then the human annotators are asked to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions.", "### Supported Tasks and Leaderboards\n\n\nWe use perplexity and BLEU score to evaluate the quality of the language generated by the model. We also compare the dialogue state generated by the model s and the ground truth state s0. Two categories of the metrics are used: exact match scores and scaled scores\n\n\nThe inference competition & leaderboard can be found here:\nURL", "### Languages\n\n\nThe text in the dataset is in English. The BCP 47 code is 'en'\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThe data is provided in two set of files. The first one has the dialogues ('air\\_dialogue\\_data') and the knowledge-base ('air\\_dialogue\\_kb')\n\n\nBuilderConfig: 'air\\_dialogue\\_data'\n\n\nBuilderConfig: 'air\\_dialogue\\_kb'", "### Data Fields\n\n\nBuilderConfig: 'air\\_dialogue\\_data':\nProvides for customer context, dialogue states and environment\n\n\n\nBuilderConfig: 'air\\_dialogue\\_kb':\nProvides for the Agent Context *ca* = (*db*, *r* )", "### Data Splits\n\n\nData is split into Train/Dev & Test in the ration of 80%, 10% and 10%\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process\n\n\nTo collect this dataset, we create a contextgenerator which provides travel and flight restrictions. We then ask human annotators to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. Key to our environment is the ease of evaluating the success of the dialogue, which is achieved by using ground-truth states (e.g., the flight being booked) generated by the restrictions. Any dialogue agent that does not generate the correct states is considered to fail.", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nNo personal and sensitive information is stored\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nAirDialogue team\n\n\nFor issues regarding HuggingFace Dataset Hub implementation Aakash Gupta", "### Licensing Information\n\n\ncc-by-nc-4.0\n\n\n@inproceedings{wei-etal-2018-airdialogue,\ntitle = \"{A}ir{D}ialogue: An Environment for Goal-Oriented Dialogue Research\",\nauthor = \"Wei, Wei and\nLe, Quoc and\nDai, Andrew and\nLi, Jia\",\nbooktitle = \"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing\",\nmonth = oct # \"-\" # nov,\nyear = \"2018\",\naddress = \"Brussels, Belgium\",\npublisher = \"Association for Computational Linguistics\",\nurl = \"URL\ndoi = \"10.18653/v1/D18-1419\",\npages = \"3844--3854\",\nabstract = \"Recent progress in dialogue generation has inspired a number of studies on dialogue systems that are capable of accomplishing tasks through natural language interactions. A promising direction among these studies is the use of reinforcement learning techniques, such as self-play, for training dialogue agents. However, current datasets are limited in size, and the environment for training agents and evaluating progress is relatively unsophisticated. We present AirDialogue, a large dataset that contains 301,427 goal-oriented conversations. To collect this dataset, we create a context-generator which provides travel and flight restrictions. We then ask human annotators to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. Key to our environment is the ease of evaluating the success of the dialogue, which is achieved by using ground-truth states (e.g., the flight being booked) generated by the restrictions. Any dialogue agent that does not generate the correct states is considered to fail. Our experimental results indicate that state-of-the-art dialogue models can only achieve a score of 0.17 while humans can reach a score of 0.91, which suggests significant opportunities for future improvement.\",\n}", "### Contributions\n\n\nThanks to @skyprince999 for adding this dataset." ]
[ "TAGS\n#task_categories-conversational #task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-generation #task_ids-dialogue-modeling #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-crowdsourced #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-nc-4.0 #region-us \n", "### Dataset Summary\n\n\nAirDialogue, is a large dataset that contains 402,038 goal-oriented conversations. To collect this dataset, we create a contextgenerator which provides travel and flight restrictions. Then the human annotators are asked to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions.", "### Supported Tasks and Leaderboards\n\n\nWe use perplexity and BLEU score to evaluate the quality of the language generated by the model. We also compare the dialogue state generated by the model s and the ground truth state s0. Two categories of the metrics are used: exact match scores and scaled scores\n\n\nThe inference competition & leaderboard can be found here:\nURL", "### Languages\n\n\nThe text in the dataset is in English. The BCP 47 code is 'en'\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThe data is provided in two set of files. The first one has the dialogues ('air\\_dialogue\\_data') and the knowledge-base ('air\\_dialogue\\_kb')\n\n\nBuilderConfig: 'air\\_dialogue\\_data'\n\n\nBuilderConfig: 'air\\_dialogue\\_kb'", "### Data Fields\n\n\nBuilderConfig: 'air\\_dialogue\\_data':\nProvides for customer context, dialogue states and environment\n\n\n\nBuilderConfig: 'air\\_dialogue\\_kb':\nProvides for the Agent Context *ca* = (*db*, *r* )", "### Data Splits\n\n\nData is split into Train/Dev & Test in the ration of 80%, 10% and 10%\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process\n\n\nTo collect this dataset, we create a contextgenerator which provides travel and flight restrictions. We then ask human annotators to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. Key to our environment is the ease of evaluating the success of the dialogue, which is achieved by using ground-truth states (e.g., the flight being booked) generated by the restrictions. Any dialogue agent that does not generate the correct states is considered to fail.", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nNo personal and sensitive information is stored\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nAirDialogue team\n\n\nFor issues regarding HuggingFace Dataset Hub implementation Aakash Gupta", "### Licensing Information\n\n\ncc-by-nc-4.0\n\n\n@inproceedings{wei-etal-2018-airdialogue,\ntitle = \"{A}ir{D}ialogue: An Environment for Goal-Oriented Dialogue Research\",\nauthor = \"Wei, Wei and\nLe, Quoc and\nDai, Andrew and\nLi, Jia\",\nbooktitle = \"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing\",\nmonth = oct # \"-\" # nov,\nyear = \"2018\",\naddress = \"Brussels, Belgium\",\npublisher = \"Association for Computational Linguistics\",\nurl = \"URL\ndoi = \"10.18653/v1/D18-1419\",\npages = \"3844--3854\",\nabstract = \"Recent progress in dialogue generation has inspired a number of studies on dialogue systems that are capable of accomplishing tasks through natural language interactions. A promising direction among these studies is the use of reinforcement learning techniques, such as self-play, for training dialogue agents. However, current datasets are limited in size, and the environment for training agents and evaluating progress is relatively unsophisticated. We present AirDialogue, a large dataset that contains 301,427 goal-oriented conversations. To collect this dataset, we create a context-generator which provides travel and flight restrictions. We then ask human annotators to play the role of a customer or an agent and interact with the goal of successfully booking a trip given the restrictions. Key to our environment is the ease of evaluating the success of the dialogue, which is achieved by using ground-truth states (e.g., the flight being booked) generated by the restrictions. Any dialogue agent that does not generate the correct states is considered to fail. Our experimental results indicate that state-of-the-art dialogue models can only achieve a score of 0.17 while humans can reach a score of 0.91, which suggests significant opportunities for future improvement.\",\n}", "### Contributions\n\n\nThanks to @skyprince999 for adding this dataset." ]
af3f2fa5462ac461b696cb300d66e07ad366057f
"\n# Dataset Card for Arabic Jordanian General Tweets\n\n## Table of Contents\n- [Dataset Card for A(...TRUNCATED)
ajgt_twitter_ar
["task_categories:text-classification","task_ids:sentiment-classification","annotations_creators:fou(...TRUNCATED)
2022-03-02T23:29:22+00:00
"{\"annotations_creators\": [\"found\"], \"language_creators\": [\"found\"], \"language\": [\"ar\"],(...TRUNCATED)
2024-01-09T11:58:01+00:00
[]
[ "ar" ]
"TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators(...TRUNCATED)
"Dataset Card for Arabic Jordanian General Tweets\n================================================\(...TRUNCATED)
["### Dataset Summary\n\n\nArabic Jordanian General Tweets (AJGT) Corpus consisted of 1,800 tweets a(...TRUNCATED)
["TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creator(...TRUNCATED)
71593d1379934286885c53d147bc863ffe830745
"\n# Dataset Card for [Dataset Name]\n\n## Table of Contents\n- [Dataset Description](#dataset-descr(...TRUNCATED)
allegro_reviews
["task_categories:text-classification","task_ids:sentiment-scoring","task_ids:text-scoring","annotat(...TRUNCATED)
2022-03-02T23:29:22+00:00
"{\"annotations_creators\": [\"found\"], \"language_creators\": [\"found\"], \"language\": [\"pl\"],(...TRUNCATED)
2024-01-09T11:59:39+00:00
[]
[ "pl" ]
"TAGS\n#task_categories-text-classification #task_ids-sentiment-scoring #task_ids-text-scoring #anno(...TRUNCATED)
"\n# Dataset Card for [Dataset Name]\n\n## Table of Contents\n- Dataset Description\n - Dataset Sum(...TRUNCATED)
["# Dataset Card for [Dataset Name]","## Table of Contents\n- Dataset Description\n - Dataset Summa(...TRUNCATED)
["TAGS\n#task_categories-text-classification #task_ids-sentiment-scoring #task_ids-text-scoring #ann(...TRUNCATED)
README.md exists but content is empty. Use the Edit dataset card button to edit it.
Downloads last month
41
Edit dataset card