Datasets:

Modalities:
Text
Formats:
json
Languages:
English
Size:
< 1K
Libraries:
Datasets
pandas
License:
YTTB-VQA / README.md
weiyueli7's picture
update YTTB-VQA v2
e86a312
metadata
task_categories:
  - visual-question-answering
language:
  - en
pretty_name: YTTB-VQA
size_categories:
  - n<1K
license: cc-by-nc-4.0

Dataset Card for Dataset Name

Dataset Description

Dataset Summary

The YTTB-VQA Dataset is a collection of 400 Youtube thumbnail question-answer pairs to evaluate the visual perception abilities of in-text images. It covers 11 categories, including technology, sports, entertainment, food, news, history, music, nature, cars, and education.

Supported Tasks and Leaderboards

This dataset supports many tasks, including visual question answering, image captioning, etc.

License

CC-By-NC-4.0

Languages

The language of the data is primarily English.

Getting Started

Creating the dataset

Run the following command to download the images and create the dataset:

python3 create_dataset.py

You will find the images in images_new and the dataset in youtube_new.json.

Dataset Structure

Data Instances

A data instance in this dataset represents entries from a collection augmented by human-generated questions submitted to BLIVA. The answer is then entered into the answer field.

Data Fields

video_id: a unique string representing a specific YouTube thumbnail image.
question: representing a human-generated question.
video_classes: representing a specific category for the YouTube thumbnail image.
answers: This represents a ground truth answer for the question made about the YouTube thumbnail image.
video link Representing the URL link for each YouTube video.

Data Splits

The data are unsplit.

Dataset Creation

Source Data

Initial Data Collection and Normalization

We randomly selected YouTube videos with text-rich thumbnails from different categories during the data collection. We recorded the unique video ID for each YouTube video and obtained the high-resolution thumbnail from the URL ”http://img.youtube.com/vi/YouTube-Video-ID/maxresdefault.jpg”.

Annotations

Annotation process

We created the annotation file with the following fields: ”video id,” question,” video classes,” answers,” and ”video link" in JSON format.

Considerations for Using the Data

Discussion of Biases

Although our dataset spans 11 categories, the ratio within each category varies. For example, 18% of the dataset pertains to education, while only 2% is dedicated to news.

Acknowledgments

The youtube thumbnails dataset is purely for academic research and not for any monetary uses. For any of the authors who saw our dataset and found their thumbnail images used inappropriately, please get in touch with us directly by this email at w1hu@ucsd.edu and we will remove the image immediately.