Datasets:
File size: 7,545 Bytes
b8665cd 627bfe6 0168dc1 627bfe6 b8665cd 591ba0e 5a9394c 3a9b676 3f290f8 3a9b676 da5c9a1 556e9ce 3f290f8 a1b3168 5324df9 a1b3168 2dee5ba 62b5fa1 6bcee00 2dee5ba 62b5fa1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 |
---
annotations_creators:
- expert-generated
language_creators:
- expert-generated
- machine-generated
languages:
- ru
licenses:
- afl-3.0
multilinguality:
- monolingual
pretty_name: 'The Pages of Early Soviet Performance (PESP) uses machine learning to
generate multiple datasets of early-Soviet illustrated periodicals related to the
performing arts. By using computer vision techniques and training a YOLO (You Only
Look Once) real-time object detection model, we are producing textual and image
data that will facilitate new avenues of research about Soviet culture during the
first decades after the October Revolution (1917-1932).
Our starting point is Princeton University Library''s Digital PUL (DPUL) where ten
titles - totaling 526 issues and approximately 26,000 pages - of Soviet performance
journals have been digitized and can be freely viewed online. Journals are a diverse
and complex genre: taken together, this collection contains hundreds of thousands
of articles, poems, editorial commentary, advertisements as well as images, illustrations
and graphic art. Today, researchers can browse the journals and view and download
high-quality page images on DPUL.'
size_categories: []
source_datasets:
- original
task_categories:
- other
task_ids: []
---
# Pages of Early Soviet Performance (PESP)
This dataset was created as part of the [Pages of Early Soviet Performance](https://cdh.princeton.edu/projects/pages-early-soviet-performance/) project at Princeton and provides text and image research data from a previously scanned [collection of illustrated periodicals](https://dpul.princeton.edu/slavic/catalog?f%5Breadonly_collections_ssim%5D%5B%5D=Russian+Illustrated+Periodicals) held by Princeton University's Slavic Collections. The project was a partnership with ITMO University in Saint Petersburg. Our work focused on document segmentation and the prediction of image, text, title, and mixedtext regions in the document images. The mixedtext category refers to segments where the typeface and text layout are mixed with other visual elements such as graphics, photographs, and illustrations. This category identifies sections that present problems for OCR and also highlights the experimental use of text, images, and other elements in the documents.
For each of the ten journals of interest in Princeton's digital collections (DPUL), we started with the IIIF manifest URI. With these manifests, we downloaded each of the 24,000 document images. The URI for each of the images is included in the dataset.
## Authors
Natalia Ermolaev, Thomas Keenan, Katherine Reischl, Andrew Janco, Quinn Dombrowski, Antonina Puchkovskaia, Alexander Jacobson, Anastasiia Mamonova, Michael Galperin and Vladislav Tretyak
## Journal manifests
- [Эрмитаж](https://figgy.princeton.edu/concern/scanned_resources/6b561fbb-ba28-4afb-91d2-d77b8728d7d9/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/6b561fbb-ba28-4afb-91d2-d77b8728d7d9/manifest)
- [Вестник искусств](https://figgy.princeton.edu/concern/scanned_resources/ad256b35-9ad0-4f75-bf83-3bad1a7c6018/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/ad256b35-9ad0-4f75-bf83-3bad1a7c6018/manifest)
- [Советский театр](https://figgy.princeton.edu/concern/scanned_resources/f33993bb-a041-40a1-b11f-f660da825583/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/f33993bb-a041-40a1-b11f-f660da825583/manifest)
- [Рабис](https://figgy.princeton.edu/concern/scanned_resources/01f4236f-0a2f-473c-946f-d9bbec12f8ea/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/01f4236f-0a2f-473c-946f-d9bbec12f8ea/manifest)
- [Даёшь](https://figgy.princeton.edu/concern/scanned_resources/e036a5da-97a8-4041-ad62-a57af44359e2/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/e036a5da-97a8-4041-ad62-a57af44359e2/manifest)
- [Персимфанс](https://figgy.princeton.edu/concern/scanned_resources/af43d19a-3659-4dd0-a0fc-4c74ce521ad6/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/af43d19a-3659-4dd0-a0fc-4c74ce521ad6/manifest)
- [Тридцать дней](https://figgy.princeton.edu/concern/scanned_resources/d2d488af-2980-4554-a9ef-aacbaf463ec8/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/d2d488af-2980-4554-a9ef-aacbaf463ec8/manifest)
- [За пролетарское искусство](https://figgy.princeton.edu/concern/scanned_resources/38f89d57-8e64-4033-97d6-b925c407584a/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/38f89d57-8e64-4033-97d6-b925c407584a/manifest)
- [Бригада художников](https://figgy.princeton.edu/concern/scanned_resources/66d00a87-5ea9-439a-a909-95d697401a2b/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/66d00a87-5ea9-439a-a909-95d697401a2b/manifest)
- [Зрелища](https://figgy.princeton.edu/concern/scanned_resources/1af8b322-a0b1-46af-8541-5c3054af8098/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/1af8b322-a0b1-46af-8541-5c3054af8098/manifest)
## Model
Using [makesense.ai](https://www.makesense.ai/) and a custom active learning application called ["Mayakovsky"](https://github.com/CDH-ITMO-Periodicals-Project/mayakovsky) we generated training data for a [YOLOv5 model](https://docs.ultralytics.com/tutorials/train-custom-datasets/). The model was fine-tuned on the new labels and predictions were generated for all images in the collection.
## OCR
Using the model's predictions for image, title, text and mixedtext segments, we cropped the image using the bounding boxes and ran OCR on each document segment using Tesseract, Google Vision, and ABBYY FineReader. Given that the output of these various OCR engines can be difficult to compare, the document segments give a common denominator for comparison of OCR outputs. Having three variations of the extracted text can be useful for experiments with OCR post-correction.
## Dataset
The dataset contains an entry for each image with the following fields:
- filename: the image name (ex. 'Советский театр_1932 No. 4_16') with journal name, year, issue, page.
- dpul: the URL for the image's journal in Digital Princeton University Library
- journal: the journal name
- year: the year of the journal issue
- issue: the issue for the image
- URI: the IIIF URI used to fetch the image from Princeton's IIIF server
- yolo: the raw model prediction (ex '3 0.1655 0.501396 0.311'), in Yolo's normalized xywh format (<object-class> <x> <y> <width> <height>). The labels are 'image'=0, 'mixedtext'=1, 'title'=2, 'textblock'=3.
- yolo_predictions: a List with a dictionary for each of the model's predictions with fields for:
- label: the predicted label
- x: the x-value location of the center point of the prediction
- y: the y-value location of the center point of the prediction
- w: the total width of the prediction's bounding box
- h: the total height of the prediction's bounding box
- abbyy_text: the text extracted from the predicted document segment using ABBY FineReader. Note that due to costs, only about 800 images have this data
- tesseract_text: the text extracted from the predicted document segment using Tesseract.
- vision_text: the text extracted from the predicted document segment using Google Vision.
# Useage
```python
from datasets import load_dataset
dataset = load_dataset('ajanco/pesp')
``` |