Datasets:

Modalities:
Image
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas
License:
VISION_LANGUAGE / README.md
neelsj's picture
Update README.md
8fbf97b verified
---
license: cdla-permissive-2.0
dataset_info:
- config_name: maze
features:
- name: id
dtype: int32
- name: image
dtype: image
- name: prompt
dtype: string
- name: ground_truth
dtype: string
- name: task
dtype: string
- name: question_type
dtype: string
- name: target_options
dtype: string
- config_name: maze_text_only
features:
- name: id
dtype: int32
- name: prompt
dtype: string
- name: ground_truth
dtype: string
- name: task
dtype: string
- name: question_type
dtype: string
- name: target_options
dtype: string
- config_name: spatial_grid
features:
- name: id
dtype: int32
- name: image
dtype: image
- name: prompt
dtype: string
- name: ground_truth
dtype: string
- name: task
dtype: string
- name: question_type
dtype: string
- name: target_options
dtype: string
- config_name: spatial_grid_text_only
features:
- name: id
dtype: int32
- name: prompt
dtype: string
- name: ground_truth
dtype: string
- name: task
dtype: string
- name: question_type
dtype: string
- name: target_options
dtype: string
- config_name: spatial_map
features:
- name: id
dtype: int32
- name: image
dtype: image
- name: prompt
dtype: string
- name: ground_truth
dtype: string
- name: task
dtype: string
- name: question_type
dtype: string
- name: target_options
dtype: string
- config_name: spatial_map_text_only
features:
- name: id
dtype: int32
- name: prompt
dtype: string
- name: ground_truth
dtype: string
- name: task
dtype: string
- name: question_type
dtype: string
- name: target_options
dtype: string
configs:
- config_name: maze
data_files:
- split: val
path: maze/maze_val.parquet
- config_name: maze_text_only
data_files:
- split: val
path: maze/maze_text_only_val.parquet
- config_name: spatial_grid
data_files:
- split: val
path: spatial_grid/spatial_grid_val.parquet
- config_name: spatial_grid_text_only
data_files:
- split: val
path: spatial_grid/spatial_grid_text_only_val.parquet
- config_name: spatial_map
data_files:
- split: val
path: spatial_map/spatial_map_val.parquet
- config_name: spatial_map_text_only
data_files:
- split: val
path: spatial_map/spatial_map_text_only_val.parquet
---
A key question for understanding multimodal vs. language capabilities of models is what is
the relative strength of the spatial reasoning and understanding in each modality, as spatial understanding is
expected to be a strength for multimodality? To test this we created a procedurally generatable, synthetic dataset
to testing spatial reasoning, navigation, and counting. These datasets are challenging and by
being procedurally generated new versions can easily be created to combat the effects of models being trained
on this data and the results being due to memorization. For each task, each question has an image and a text
representation that is sufficient for answering each question.
This dataset has three tasks that test: Spatial Understanding (Spatial-Map), Nav-
igation (Maze), and Counting (Spatial-Grid). Each task has three conditions, with respect to the input
modality, 1) text-only, input and a question, 2) vision-only, which is the standard task of visual-question an-
swering that consists of a vision-only input and a question, and 3) vision-text includes both text and image
representations with the question. Each condition includes 1500
images and text pairs for a total of 4500.
__Spatial Map__
The dataset consists of spatial relationships for random layouts of symbolic objects with text names on white background.
Each object is associated with a unique location name, such as Unicorn Umbrellas and Gale Gifts. To study the impact of modality,
the textual representation of each input consists of pairwise relations such as Brews Brothers Pub
is to the Southeast of Whale’s Watches. The questions include asking about the spatial
relationships between two locations and the number of objects that meet specific spatial criteria.
The dataset includes 3 conditions: text only, image only, and text+image. Each condition includes 1500 images and text pairs for a total of 4500.
There are 3 question types:
1) In which direction is one object to another (answer is a direction)
2) Which object is to the direction of another (answer is an object name)
3) How many objects are in a direction of another (answer is a number)
Each question is multiple choice.
__Maze__
The dataset consists of small mazes with questions asked about the maze. Each sample can be
represented as colored blocks where different colors signify distinct elements: a green block marks
the starting point (S), a red block indicates the exit (E), black blocks represent impassable walls,
white blocks denote navigable paths, and blue blocks trace the path from S to E. The objective is to
navigate from S to E following the blue path, with movement permitted in the four cardinal directions
(up, down, left, right). Alternatively, each input can be depicted in textual format using ASCII code.
The questions asked include counting the number of turns from S to E and determining the spatial relationship
between S and E.
The dataset includes 3 conditions: text only, image only, and text+image. Each condition includes 1500 images and text pairs for a total of 4500.
There are 3 question types:
1) How many right turns on the path from start to end (answer is a number)
2) How many total turns on the path from start to end (answer is a number)
3) Where is the exit releative to the start (answer is direction or yes/no)
Each question is multiple choice.
__Spatial Grid__
Each input consists of a grid of cells, each containing an image (e.g.,a rabbit). Alternatively, this grid
can also be represented in a purely textual format; for instance, the first row might be described as:
elephant | cat | giraffe | elephant | cat. The evaluations focus on tasks such as counting specific objects (e.g., rabbits) and
identifying the object located at a specific coordinate in the grid (e.g., first row, second column).
The dataset includes 3 conditions: text only, image only, and text+image. Each condition includes 1500 images and text pairs for a total of 4500 questions.
There are 3 question types:
1) How many blocks contain a specific animal (answer is a number)
2) What animal is in one specific block, adressed by top-left, top, right, etc. (answer is an animal name)
3) What animal is in one specific block, addressed by row, column (answer is an animal name)
Each question is multiple choice.
---
More details here: https://arxiv.org/pdf/2406.14852