|
--- |
|
language: |
|
- en |
|
paperswithcode_id: wiki-40b |
|
pretty_name: Wiki-40B |
|
dataset_info: |
|
- config_name: ar |
|
features: |
|
- name: wikidata_id |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
- name: version_id |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 773508885 |
|
num_examples: 220885 |
|
- name: validation |
|
num_bytes: 44102674 |
|
num_examples: 12198 |
|
- name: test |
|
num_bytes: 43755879 |
|
num_examples: 12271 |
|
download_size: 413683528 |
|
dataset_size: 861367438 |
|
- config_name: de |
|
features: |
|
- name: wikidata_id |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
- name: version_id |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 4988094268 |
|
num_examples: 1554910 |
|
- name: validation |
|
num_bytes: 278101948 |
|
num_examples: 86068 |
|
- name: test |
|
num_bytes: 278024815 |
|
num_examples: 86594 |
|
download_size: 3174352286 |
|
dataset_size: 5544221031 |
|
- config_name: en |
|
features: |
|
- name: wikidata_id |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
- name: version_id |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 9423468036 |
|
num_examples: 2926536 |
|
- name: validation |
|
num_bytes: 527374301 |
|
num_examples: 163597 |
|
- name: test |
|
num_bytes: 522210646 |
|
num_examples: 162274 |
|
download_size: 6183831905 |
|
dataset_size: 10473052983 |
|
- config_name: fr |
|
features: |
|
- name: wikidata_id |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
- name: version_id |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 3850031120 |
|
num_examples: 1227206 |
|
- name: validation |
|
num_bytes: 216405364 |
|
num_examples: 68655 |
|
- name: test |
|
num_bytes: 215243874 |
|
num_examples: 68004 |
|
download_size: 2246390244 |
|
dataset_size: 4281680358 |
|
- config_name: it |
|
features: |
|
- name: wikidata_id |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
- name: version_id |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 1998187938 |
|
num_examples: 732609 |
|
- name: validation |
|
num_bytes: 109399796 |
|
num_examples: 40684 |
|
- name: test |
|
num_bytes: 108160871 |
|
num_examples: 40443 |
|
download_size: 1330554944 |
|
dataset_size: 2215748605 |
|
- config_name: ja |
|
features: |
|
- name: wikidata_id |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
- name: version_id |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 7719156890 |
|
num_examples: 745392 |
|
- name: validation |
|
num_bytes: 423396781 |
|
num_examples: 41576 |
|
- name: test |
|
num_bytes: 424775191 |
|
num_examples: 41268 |
|
download_size: 2914923230 |
|
dataset_size: 8567328862 |
|
- config_name: nl |
|
features: |
|
- name: wikidata_id |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
- name: version_id |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 906908479 |
|
num_examples: 447555 |
|
- name: validation |
|
num_bytes: 51519150 |
|
num_examples: 25201 |
|
- name: test |
|
num_bytes: 49492508 |
|
num_examples: 24776 |
|
download_size: 594312303 |
|
dataset_size: 1007920137 |
|
- config_name: zh-cn |
|
features: |
|
- name: wikidata_id |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
- name: version_id |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 902812807 |
|
num_examples: 549672 |
|
- name: validation |
|
num_bytes: 50487729 |
|
num_examples: 30299 |
|
- name: test |
|
num_bytes: 49584239 |
|
num_examples: 30355 |
|
download_size: 667605463 |
|
dataset_size: 1002884775 |
|
- config_name: zh-tw |
|
features: |
|
- name: wikidata_id |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
- name: version_id |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 3254625339 |
|
num_examples: 552031 |
|
- name: validation |
|
num_bytes: 185024571 |
|
num_examples: 30739 |
|
- name: test |
|
num_bytes: 181148137 |
|
num_examples: 30670 |
|
download_size: 1375185673 |
|
dataset_size: 3620798047 |
|
configs: |
|
- config_name: ar |
|
data_files: |
|
- split: train |
|
path: ar/train-* |
|
- split: validation |
|
path: ar/validation-* |
|
- split: test |
|
path: ar/test-* |
|
- config_name: de |
|
data_files: |
|
- split: train |
|
path: de/train-* |
|
- split: validation |
|
path: de/validation-* |
|
- split: test |
|
path: de/test-* |
|
- config_name: en |
|
data_files: |
|
- split: train |
|
path: en/train-* |
|
- split: validation |
|
path: en/validation-* |
|
- split: test |
|
path: en/test-* |
|
- config_name: fr |
|
data_files: |
|
- split: train |
|
path: fr/train-* |
|
- split: validation |
|
path: fr/validation-* |
|
- split: test |
|
path: fr/test-* |
|
- config_name: it |
|
data_files: |
|
- split: train |
|
path: it/train-* |
|
- split: validation |
|
path: it/validation-* |
|
- split: test |
|
path: it/test-* |
|
- config_name: ja |
|
data_files: |
|
- split: train |
|
path: ja/train-* |
|
- split: validation |
|
path: ja/validation-* |
|
- split: test |
|
path: ja/test-* |
|
- config_name: nl |
|
data_files: |
|
- split: train |
|
path: nl/train-* |
|
- split: validation |
|
path: nl/validation-* |
|
- split: test |
|
path: nl/test-* |
|
- config_name: zh-cn |
|
data_files: |
|
- split: train |
|
path: zh-cn/train-* |
|
- split: validation |
|
path: zh-cn/validation-* |
|
- split: test |
|
path: zh-cn/test-* |
|
- config_name: zh-tw |
|
data_files: |
|
- split: train |
|
path: zh-tw/train-* |
|
- split: validation |
|
path: zh-tw/validation-* |
|
- split: test |
|
path: zh-tw/test-* |
|
--- |
|
|
|
# Dataset Card for "wiki40b" |
|
|
|
## Table of Contents |
|
- [Dataset Description](#dataset-description) |
|
- [Dataset Summary](#dataset-summary) |
|
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) |
|
- [Languages](#languages) |
|
- [Dataset Structure](#dataset-structure) |
|
- [Data Instances](#data-instances) |
|
- [Data Fields](#data-fields) |
|
- [Data Splits](#data-splits) |
|
- [Dataset Creation](#dataset-creation) |
|
- [Curation Rationale](#curation-rationale) |
|
- [Source Data](#source-data) |
|
- [Annotations](#annotations) |
|
- [Personal and Sensitive Information](#personal-and-sensitive-information) |
|
- [Considerations for Using the Data](#considerations-for-using-the-data) |
|
- [Social Impact of Dataset](#social-impact-of-dataset) |
|
- [Discussion of Biases](#discussion-of-biases) |
|
- [Other Known Limitations](#other-known-limitations) |
|
- [Additional Information](#additional-information) |
|
- [Dataset Curators](#dataset-curators) |
|
- [Licensing Information](#licensing-information) |
|
- [Citation Information](#citation-information) |
|
- [Contributions](#contributions) |
|
|
|
## Dataset Description |
|
|
|
- **Homepage:** [https://research.google/pubs/pub49029/](https://research.google/pubs/pub49029/) |
|
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
- **Size of downloaded dataset files:** 0.00 MB |
|
- **Size of the generated dataset:** 10.47 GB |
|
- **Total amount of disk used:** 10.47 GB |
|
|
|
### Dataset Summary |
|
|
|
Clean-up text for 40+ Wikipedia languages editions of pages |
|
correspond to entities. The datasets have train/dev/test splits per language. |
|
The dataset is cleaned up by page filtering to remove disambiguation pages, |
|
redirect pages, deleted pages, and non-entity pages. Each example contains the |
|
wikidata id of the entity, and the full Wikipedia article after page processing |
|
that removes non-content sections and structured objects. |
|
|
|
### Supported Tasks and Leaderboards |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
### Languages |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
## Dataset Structure |
|
|
|
### Data Instances |
|
|
|
#### en |
|
|
|
- **Size of downloaded dataset files:** 0.00 MB |
|
- **Size of the generated dataset:** 10.47 GB |
|
- **Total amount of disk used:** 10.47 GB |
|
|
|
An example of 'train' looks as follows. |
|
``` |
|
|
|
``` |
|
|
|
### Data Fields |
|
|
|
The data fields are the same among all splits. |
|
|
|
#### en |
|
- `wikidata_id`: a `string` feature. |
|
- `text`: a `string` feature. |
|
- `version_id`: a `string` feature. |
|
|
|
### Data Splits |
|
|
|
|name| train |validation| test | |
|
|----|------:|---------:|-----:| |
|
|en |2926536| 163597|162274| |
|
|
|
## Dataset Creation |
|
|
|
### Curation Rationale |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
### Source Data |
|
|
|
#### Initial Data Collection and Normalization |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
#### Who are the source language producers? |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
### Annotations |
|
|
|
#### Annotation process |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
#### Who are the annotators? |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
### Personal and Sensitive Information |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
## Considerations for Using the Data |
|
|
|
### Social Impact of Dataset |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
### Discussion of Biases |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
### Other Known Limitations |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
## Additional Information |
|
|
|
### Dataset Curators |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
### Licensing Information |
|
|
|
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
|
|
|
### Citation Information |
|
|
|
``` |
|
|
|
``` |
|
|
|
|
|
### Contributions |
|
|
|
Thanks to [@jplu](https://github.com/jplu), [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf), [@albertvillanova](https://github.com/albertvillanova), [@lhoestq](https://github.com/lhoestq) for adding this dataset. |