Datasets:

Modalities:
Image
Text
Formats:
json
ArXiv:
Libraries:
Datasets
Dask
License:
File size: 12,108 Bytes
593e53f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
72e8b12
593e53f
72e8b12
593e53f
72e8b12
593e53f
a2f1b52
 
 
72e8b12
a2f1b52
02d776e
 
 
 
a2f1b52
 
593e53f
e2c8e50
 
 
a018b7f
 
e2c8e50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a018b7f
e2c8e50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a018b7f
 
 
a34995c
 
 
 
 
a018b7f
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
---
license: cc-by-sa-4.0

language:
  - eng
  - ind
  - ind
  - zho
  - kor
  - kor
  - jpn
  - jpn
  - sun
  - jav
  - jav
  - ces
  - spa
  - fra
  - ara
  - hin
  - ben
  - mar
  - sin
  - yor
  - yue
  - nan
  - nan
  - tgl
  - tha
  - aze
  - rus
  - rus
  - ita
  - srd
multilinguality:
  - multilingual
language_details: >-
  en, id_formal, id_casual, zh_cn, ko_formal, ko_casual, ja_formal, ja_casual,
  su_loma, jv_krama, jv_ngoko, cs, es, fr, ar, hi, bn, mr, si_formal_spoken, yo,
  yue, nan, nan_spoken, tl, th, az, ru_formal, ru_casual, it, sc
configs:
  - config_name: task1
    data_files:
      - split: test_large
        path: hf_prompt/large_eval_task1/*
      - split: test_small
        path: hf_prompt/small_eval_task1/*
      - split: train
        path: hf_prompt/train_task1/*
        
  - config_name: task2
    data_files:
      - split: test_large
        path: hf_prompt/large_eval_task2/*
      - split: test_small
        path: hf_prompt/small_eval_task2/*
      - split: train
        path: hf_prompt/train_task2/*
---

# WorldCuisines: A Massive-Scale Benchmark for Multilingual and Multicultural Visual Question Answering on Global Cuisines

![Overview](./images/tasks.png)

WorldCuisines is a massive-scale visual question answering (VQA) benchmark for multilingual and multicultural understanding through global cuisines. The dataset contains text-image pairs across 30 languages and dialects, spanning 9 language families and featuring over 1 million data points, making it the largest multicultural VQA benchmark as of 17 October 2024.

## Overview
We develop both a VQA dataset (**WC-VQA**) and a curated Knowledge Base (KB) for world cuisines (**WC-KB**). The **WC-VQA** dataset is constructed using **WC-KB**, which serves as the primary data source. We design two tasks as follows:

#### Task 1: Dish Name Prediction
This task requires predicting the name of a dish based on its image, a question, and contextual information. It is divided into three subtasks, each with a distinct query type:
- (a) **No-context question**: Predict the dish name without additional context.
- (b) **Contextualized question**: Predict the dish name with provided contextual information.
- (c) **Adversarial contextualized question**: Predict the dish name with misleading or adversarial context.
#### Task 2: Location Prediction
This task involves predicting the location where the dish is commonly consumed or originated, based on the dish image, a question, and contextual information.

**WC-KB** encompasses 2,414 dishes worldwide, including 6,045 images and metadata, covering both coarse-grained (e.g., stew) and fine-grained categories (e.g., beef stew), locations, and regional cuisines. It also features multilingual translations of 90 crowd-sourced prompt templates and 401 parallel data entries (i.e., multilingual information) for location and regional cuisine information.
From **WC-KB**, we construct **WC-VQA**, a multilingual parallel VQA dataset with 1 million samples encompassing over 30 languages and dialects, including various varieties and registers, such as formal and casual styles, with high-quality human annotations. The VQA is designed to evaluate models' ability to understand cultural food names and their origins.

We provide **WC-VQA** evaluation datasets in two sizes (12,000 and 60,000 instances) alongside a training dataset (1,080,000 instances).
The table below provides more detailed statistics regarding the number of VQA instances and images for each data split.

![StatisticsTable](./images/stats_table.png)

## Dataset Construction

Our data sources are gathered from [Wikipedia](https://wikipedia.org) and [Wikimedia Commons](https://commons.wikimedia.org) to ensure they can be easily redistributed under an accepted open-source license. The data construction process involves four key steps: 
1. Dish selection
2. Metadata annotation
3. Quality assurance
4. Data compilation.

### Dish Selection

We compile a comprehensive list of dish names sourced from Wikipedia. We manually review pages that feature lists of dishes to determine whether each dish is a specialty unique to a specific culture, as we aim to focus on dishes that have distinct cultural significance. We exclude generic categories, such as ice cream, which lacks a specific cultural association. We ensure that each dish on our list has its own dedicated Wikipedia page. If a dish does not have a Wikipedia page, it is also excluded from our compilation. This meticulous approach ensures that our dataset is both culturally relevant and well-documented.

### Metadata Annotation

Given a dish name and its corresponding Wikipedia page link, annotators manually compile metadata based on the provided information. This metadata includes:
- **Visual Representation**: Images sourced from Wikimedia Commons are included, along with their license information.
- **Categorization**: Dishes are classified into both coarse-grained (e.g., rice, bread) and fine-grained (e.g., fried rice, flatbread) categories.
- **Description**: Annotators provide a description of each dish based on the content from its Wikipedia page, avoiding the use of the dish's name, origin, or any distinctive keywords that uniquely identify the dish.
- **Cuisine**: The dish's origin cuisine and any cuisines with which it is strongly associated.
- **Geographic Distribution**: This includes the dish's associated countries, area (city or region), and broader continental region.

### Quality Assurance

Before beginning the quality assurance process, we first identify common issues that arise during annotation and develop automated rules to detect easily identifiable annotation errors, such as incorrect string formatting. Annotators are then asked to correct these errors. To further ensure data quality and validity, we conduct several rounds of quality assurance:

1. **Image Quality**: We remove instances where images are blurry, dark, or contain distracting elements such as people or other dishes. We also verify image licenses by cross-referencing them with information on Wikimedia Commons.
2. **Categorization and Descriptions**: We refine dish categorization and descriptions, ensuring consistency in category assignments and keeping descriptions free from "information breaches" (e.g., excluding regional details from the description).
3. **Cuisine Names and Geographic Accuracy**: We standardize cuisine names and meticulously review all country and area information for accuracy.

This comprehensive approach guarantees the integrity and reliability of our dataset.

### Data Compilation

In this phase, we verify the overall quality check done by annotators, identifying any potential inconsistencies missed during quality assurance. We then compile the dataset by collecting the metadata into a single file.

## VQA Generation

In this phase, we generate VQA data by sampling from **WC-KB**. A VQA data entry comprises a visual image, question text, and answer text. This process involves four stages:

1. Conducting a similarity search for dish names
2. Constructing questions and contexts
3. Translating these elements into multiple languages
4. Generating the VQA triplets.

### Dish Names Similarity Search

To identify similar dishes in our dataset, we follow the approach from Winata et al. (2024) to employ a multilingual model E5$_\text{LARGE}$ Instruct (Wang et al. 2024) for computing text embeddings. Formally, given a dish $x$ with name $x_{\text{name}}$ and text description $x_{\text{desc}}$, we use a multilingual model $\theta$ to compute the embedding vector $v_x = \theta(\{x_\text{name};x_\text{desc}\})$, and then apply cosine similarity to compute a score $s = \text{similarity}(v_i, v_j)$ between dish $i$ and dish $j$. For each dish, we consider the top-$k$ most similar dishes to generate distractors in the multiple choice questions.

### Question and Context Construction

Dish name prediction (Task 1) is divided into three question variations depending on the context:

1. **No-context question**: Simply asks for the name of the dish without any provided context.
2. **Contextualized question**: Provides additional information related to cuisine or location.
3. **Adversarial contextualized question**: Similar to contextualized questions, but may include misleading location information to assess the model's robustness to irrelevant details.

For regional cuisine prediction (Task 2), only a basic question without any provided context is available.

### Multiple Language Translation

#### Question and Context

All questions and contexts are initially collected in English and are then carefully translated by native speakers into 30 language varieties: 23 different languages, with 7 languages having two varieties. Translators prioritize naturalness, followed by diversity in translations when duplication occurs.

#### Food Name Alias

Using Wikipedia pages as our primary source, we verify if the English page has translations available in other languages. This enables us to extract dish names in multiple languages and compile them as translations for each dish. We utilize both the Wikipedia page titles in various languages and alias text from the English page. These translations enhance cultural relevance and accuracy for multilingual prompts. When translation is unavailable, we use the English name as the default.

#### Locations and Cuisines

With over 400 unique locations, including countries, cities, and areas, we first translate the English locations into other languages using GPT-4, followed by proofreading by native speakers. Regional cuisine names (the adjective form of the location in English) are translated in the same manner.

#### Morphological Inflections

In languages with rich inflectional morphology (e.g., Czech or Spanish), words are modified to express different grammatical categories (e.g., number, gender, or case). We provide a framework for human translators to use natural inflections in the prompt template while keeping the inflections as few as possible.

### Generating VQA Triplets

To ensure no overlap between training and testing subsets, we split the dishes and multilingual questions into two subsets. For each subset, we randomly sample dish and question pairs. We use the dish entry in our KB dataset to select the image and inject the location into the context, if applicable. Answer candidates for multiple-choice questions are picked using the similarity search process. We repeat this process until the desired number of training or test samples is reached, discarding duplicates.

## Ethical Considerations

Our research focuses on evaluating VLMs within the context of multilingual and multicultural VQA, a field that holds significant implications for diverse multilingual communities. We are committed to conducting our data collection and evaluations with the highest standards of transparency and fairness. To achieve this, we have adopted a crowd-sourcing approach for the annotation process, inviting volunteers to contribute and become co-authors if they provide significant contributions. We follow the guidelines from ACL for authorship eligibility as shown in  https://www.aclweb.org/adminwiki/index.php/Authorship_Changes_Policy_for_ACL_Conference_Papers. In line with our commitment to openness and collaboration, we will release our dataset under an open-source license, CC-BY-SA 4.0.

## Contact

E-mail: [Genta Indra Winata](genta.winata@capitalone.com) and [Frederikus Hudi](frederikus.hudi.fe7@is.naist.jp)

## Citation

If you find this dataset useful, please cite the following works

```bibtex
@article{winata2024worldcuisines,
  title={WorldCuisines: A Massive-Scale Benchmark for Multilingual and Multicultural Visual Question Answering on Global Cuisines},
  author={Winata, Genta Indra and Hudi, Frederikus and Irawan, Patrick Amadeus and Anugraha, David and Putri, Rifki Afina and Wang, Yutong and Nohejl, Adam and Prathama, Ubaidillah Ariq and Ousidhoum, Nedjma and Amriani, Afifa and others},
  journal={arXiv preprint arXiv:2410.12705},
  year={2024}
}
```