The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Error code: DatasetGenerationError Exception: CastError Message: Couldn't cast identifier: string creator: string title: string publication_date: int64 word_count: int64 text: string __index_level_0__: int64 -- schema metadata -- pandas: '{"index_columns": ["__index_level_0__"], "column_indexes": [{"na' + 1027 to {'identifier': Value(dtype='string', id=None), 'creator': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'publication_date': Value(dtype='int64', id=None), 'word_count': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)} because column names don't match Traceback: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1492, in compute_config_parquet_and_info_response fill_builder_info(builder, hf_endpoint=hf_endpoint, hf_token=hf_token, validate=validate) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 683, in fill_builder_info ) = retry_validate_get_features_num_examples_size_and_compression_ratio( File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 602, in retry_validate_get_features_num_examples_size_and_compression_ratio validate(pf) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 640, in validate raise TooBigRowGroupsError( worker.job_runners.config.parquet_and_info.TooBigRowGroupsError: Parquet file has too big row groups. First row group has 673579700 which exceeds the limit of 300000000 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1995, in _prepare_split_single for _, table in generator: File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 797, in wrapped for item in generator(*args, **kwargs): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/parquet/parquet.py", line 97, in _generate_tables yield f"{file_idx}_{batch_idx}", self._cast_table(pa_table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/parquet/parquet.py", line 75, in _cast_table pa_table = table_cast(pa_table, self.info.features.arrow_schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2302, in table_cast return cast_table_to_schema(table, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2256, in cast_table_to_schema raise CastError( datasets.table.CastError: Couldn't cast identifier: string creator: string title: string publication_date: int64 word_count: int64 text: string __index_level_0__: int64 -- schema metadata -- pandas: '{"index_columns": ["__index_level_0__"], "column_indexes": [{"na' + 1027 to {'identifier': Value(dtype='string', id=None), 'creator': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'publication_date': Value(dtype='int64', id=None), 'word_count': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None)} because column names don't match The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1505, in compute_config_parquet_and_info_response parquet_operations, partial, estimated_dataset_info = stream_convert_to_parquet( File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1099, in stream_convert_to_parquet builder._prepare_split( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1882, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2038, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
identifier
string | creator
string | title
string | publication_date
int64 | word_count
int64 | text
string |
---|---|---|---|---|---|
bim_early-english-books-1641-1700_the-devout-communicant-e_1675 | null | The devout communicant exemplifi'd, ... 1675 | 1,675 | 46,844 | "The DEVOUT /S 72 \nCOMMUNI CANT \nExemplified, in his Behaviour \n\n- Bekoze, At, and After \n\n\nT(...TRUNCATED) |
bim_early-english-books-1641-1700_the-best-fence-against-p_1686 | null | The best fence against popery:... 1686 | 1,686 | 24,011 | "IvinDicatION \n\n\n| ; \n: \n\n\nOF THE \n\n\n| Power of wing K 1 N G \n\n\n[To which is added, Que(...TRUNCATED) |
bim_eighteenth-century_the-works-of-hugh-kelly_kelly-hugh-dramatic-wr_1778 | Kelly, Hugh, Dramatic writer. | The works of Hugh Kelly. 1778 | 1,778 | 220,835 | "„„ Bon oe f \n2 8 2 2 * . * 4 % \n„ CT T 3 * \nW205 275 PSB OR, BN WE . br, WE 2 g \n* \"= =2(...TRUNCATED) |
bim_eighteenth-century_grammar-made-familiar-an_newbery-john_1769 | Newbery, John | "Grammar made familiar and easy to young gentlemen, ladies, and foreigners, being the first volume o(...TRUNCATED) | 1,769 | 25,880 | "Young Ge Lad c \n\n\nand Forei gners, \n5 | * Being the. \n\n\n= FIRST VOLUME \nSz THE; 1 1 4 \n\n\(...TRUNCATED) |
bim_early-english-books-1641-1700_christ-and-his-saints-_jelinger-christopher_1656 | Jelinger, Christopher. | Christ and his saints. ... 1656 | 1,656 | 9,336 | "S AINTS, \n\n\n= \n\n\nSpend! theirtime togeth \nBy and Night: I \n\n\nChriſt and his Saints WEE t(...TRUNCATED) |
bim_eighteenth-century_1733 | null | "The Historical Register, containing an impartial relation of all transactions. With a chronological(...TRUNCATED) | 1,733 | 174,024 | "e r \"387 {3K TR \n\n** . \n2 I te a 4. 7 2 \n\n0 : ds age WT If ; \n\n\n5 : 8888 8 p 11 * F - TIM (...TRUNCATED) |
bim_early-english-books-1475-1640_the-light-unchangeable-_smith-richard-quaker_1677 | Smith, Richard, Quaker. | The light unchangeable: ... 1677 | 1,677 | 17,599 | "THE \n\n\n| _1ght Unchangeablc \n\n\nAND \nTruth and Good Order, juſtified \nagainſt Error and Di(...TRUNCATED) |
bim_eighteenth-century_memoirs-of-mary-a-novel_gunning-mrs-susannah_1793_4 | Gunning, Mrs. (Susannah) | Memoirs of Mary, a novel. By Mrs. Gunning. In five volumes. ... 1793: Vol 4 | 1,793 | 44,845 | "By Mrs. GUNMNIN G, \n\n\n— — \n— — \n\n\nIN FIVE VOLUMES. \nVOL. IV. \n\n\nE 1 0 IM \n\n\nP(...TRUNCATED) |
bim_eighteenth-century_a-greek-and-english-lexi_parkhurst-john_1769 | Parkhurst, John | "A Greek and English lexicon to the New Testament: ... To this work is prefixed, a plain and easy Gr(...TRUNCATED) | 1,769 | 570,313 | "\"FN. W HIC H pF” WM 4 5 15 \n\n\nI. . The WORD: 1 RASES occugring in thoſe SD Boo 5 \n| _ „ J(...TRUNCATED) |
bim_eighteenth-century_the-country-clergymans-_holmes-william_1783 | Holmes, William | "The country clergyman's advice to his parishioners, explaining what they are to believe, and do, in(...TRUNCATED) | 1,783 | 17,175 | "os . \nCOUNTRY CLERGYMAN's \n\n\nA ÞD VS 6 'Þ \nTO HIS \n\n\nPARISHIONERS, \n\n\n| EXPLAINING \n\(...TRUNCATED) |
🇬🇧 Midlle-English Public Domain 🇬🇧
Middle-English-Public Domain or Middle-English-PD is a large collection aiming to aggregate all midlle-age English monographies, periodicals and texts in the public domain. As of March 2024, it is the biggest middle-age English open corpus.
Dataset summary
The collection contains 204,472 individual titles making up 12,100,748,108 words recovered from multiple sources, including Internet Archive and various European national libraries and cultural heritage institutions. Each parquet file has the full text of 2,000 books selected at random.
Curation method
The composition of the dataset adheres to the criteria for public domain works in the EU and, consequently, all Berne-countries for EU authors: any publication whose author is dead for more than 70 years. Additionally, the initial consolidation of public domain status for cultural heritage operates in the EU under the 2019 Copyright Directive (art. 14).
As of March 2024, to limit rights verification, we have retained exclusively titles published prior to 1884.
The corpus will be expanded at a later stage to encompass late 19th century and early 20th century publications, after checking for public domain validity.
Uses
The collection aims to expand the availability of open works for the training of Large Language Models. The text can be used for model training and republished without restriction for reproducibility purposes.
The rationales for creation of this collection are multifold:
- Scientific: We observe that the closure of training corpora represents a major barrier to AI research. Large language models face a real crisis of reproducibility.
- Legal: With the adoption of the AI Act with its obligations in terms of copyright law compliance for the pretraining corpora, the European AI ecosystem will have to change its provenance practices.
- Cultural: The linguistic diversity of the European Union is currently underrepresented. Unlike web archives, open, heritage, administrative, or scientific texts are often of high quality: they are long, multilingual, and editorialized publications.
- Economical: Today, value capture is concentrated on players whose financial resources are already considerable, allowing them to collect or purchase data at a high price. Making a royalty-free corpus available to as many people as possible frees innovation in uses and minimizes economic dependencies on dominant actors.
License
The entire collection is in the public domain in all regions. This means that the patrimonial rights of each individual or collective right holders have expired.
There has been a debate for years in Europe over the definition of public domain and the possibility to restrict its use. Since 2019, the EU Copyright Directive states that "Member States shall provide that, when the term of protection of a work of visual art has expired, any material resulting from an act of reproduction of that work is not subject to copyright or related rights, unless the material resulting from that act of reproduction is original in the sense that it is the author's own intellectual creation." (art. 14)
Future work
This dataset is not a one-time work but will continue to evolve significantly in three directions:
- Expansion of the dataset to the late 19th and early 20th century works and its further enhancement with currently unexploited collections coming from European patrimonial data repositories.
- Correction of computer generated errors in the text. All the texts have been transcribed automatically through the use of Optical Character Recognition (OCR) software. The original files have been digitized over a long time period (since the mid-2000s) and some documents should be. Future versions will strive either to re-OCRize the original text or use experimental LLM models for partial OCR correction.
- Enhancement of the structure/editorial presentation of the original text. Some parts of the original documents are likely unwanted for large scale analysis or model training (header, page count…). Additionally, some advanced document structures like tables or multi-column layout are unlikely to be well-formatted.
Acknowledgements
The corpus was stored and processed with the generous support of Scaleway. It was built up with the support and concerted efforts of the state start-up LANGU:IA (start-up d’Etat), supported by the French Ministry of Culture and DINUM, as part of the prefiguration of the service offering of the Alliance for Language technologies EDIC (ALT-EDIC).
Corpus collection has been largely facilitated thanks to the open science LLM community insights and cooperation (Occiglot, Eleuther AI, Allen AI).
- Downloads last month
- 61