File size: 12,108 Bytes
593e53f 72e8b12 593e53f 72e8b12 593e53f 72e8b12 593e53f a2f1b52 72e8b12 a2f1b52 02d776e a2f1b52 593e53f e2c8e50 a018b7f e2c8e50 a018b7f e2c8e50 a018b7f a34995c a018b7f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 |
---
license: cc-by-sa-4.0
language:
- eng
- ind
- ind
- zho
- kor
- kor
- jpn
- jpn
- sun
- jav
- jav
- ces
- spa
- fra
- ara
- hin
- ben
- mar
- sin
- yor
- yue
- nan
- nan
- tgl
- tha
- aze
- rus
- rus
- ita
- srd
multilinguality:
- multilingual
language_details: >-
en, id_formal, id_casual, zh_cn, ko_formal, ko_casual, ja_formal, ja_casual,
su_loma, jv_krama, jv_ngoko, cs, es, fr, ar, hi, bn, mr, si_formal_spoken, yo,
yue, nan, nan_spoken, tl, th, az, ru_formal, ru_casual, it, sc
configs:
- config_name: task1
data_files:
- split: test_large
path: hf_prompt/large_eval_task1/*
- split: test_small
path: hf_prompt/small_eval_task1/*
- split: train
path: hf_prompt/train_task1/*
- config_name: task2
data_files:
- split: test_large
path: hf_prompt/large_eval_task2/*
- split: test_small
path: hf_prompt/small_eval_task2/*
- split: train
path: hf_prompt/train_task2/*
---
# WorldCuisines: A Massive-Scale Benchmark for Multilingual and Multicultural Visual Question Answering on Global Cuisines
![Overview](./images/tasks.png)
WorldCuisines is a massive-scale visual question answering (VQA) benchmark for multilingual and multicultural understanding through global cuisines. The dataset contains text-image pairs across 30 languages and dialects, spanning 9 language families and featuring over 1 million data points, making it the largest multicultural VQA benchmark as of 17 October 2024.
## Overview
We develop both a VQA dataset (**WC-VQA**) and a curated Knowledge Base (KB) for world cuisines (**WC-KB**). The **WC-VQA** dataset is constructed using **WC-KB**, which serves as the primary data source. We design two tasks as follows:
#### Task 1: Dish Name Prediction
This task requires predicting the name of a dish based on its image, a question, and contextual information. It is divided into three subtasks, each with a distinct query type:
- (a) **No-context question**: Predict the dish name without additional context.
- (b) **Contextualized question**: Predict the dish name with provided contextual information.
- (c) **Adversarial contextualized question**: Predict the dish name with misleading or adversarial context.
#### Task 2: Location Prediction
This task involves predicting the location where the dish is commonly consumed or originated, based on the dish image, a question, and contextual information.
**WC-KB** encompasses 2,414 dishes worldwide, including 6,045 images and metadata, covering both coarse-grained (e.g., stew) and fine-grained categories (e.g., beef stew), locations, and regional cuisines. It also features multilingual translations of 90 crowd-sourced prompt templates and 401 parallel data entries (i.e., multilingual information) for location and regional cuisine information.
From **WC-KB**, we construct **WC-VQA**, a multilingual parallel VQA dataset with 1 million samples encompassing over 30 languages and dialects, including various varieties and registers, such as formal and casual styles, with high-quality human annotations. The VQA is designed to evaluate models' ability to understand cultural food names and their origins.
We provide **WC-VQA** evaluation datasets in two sizes (12,000 and 60,000 instances) alongside a training dataset (1,080,000 instances).
The table below provides more detailed statistics regarding the number of VQA instances and images for each data split.
![StatisticsTable](./images/stats_table.png)
## Dataset Construction
Our data sources are gathered from [Wikipedia](https://wikipedia.org) and [Wikimedia Commons](https://commons.wikimedia.org) to ensure they can be easily redistributed under an accepted open-source license. The data construction process involves four key steps:
1. Dish selection
2. Metadata annotation
3. Quality assurance
4. Data compilation.
### Dish Selection
We compile a comprehensive list of dish names sourced from Wikipedia. We manually review pages that feature lists of dishes to determine whether each dish is a specialty unique to a specific culture, as we aim to focus on dishes that have distinct cultural significance. We exclude generic categories, such as ice cream, which lacks a specific cultural association. We ensure that each dish on our list has its own dedicated Wikipedia page. If a dish does not have a Wikipedia page, it is also excluded from our compilation. This meticulous approach ensures that our dataset is both culturally relevant and well-documented.
### Metadata Annotation
Given a dish name and its corresponding Wikipedia page link, annotators manually compile metadata based on the provided information. This metadata includes:
- **Visual Representation**: Images sourced from Wikimedia Commons are included, along with their license information.
- **Categorization**: Dishes are classified into both coarse-grained (e.g., rice, bread) and fine-grained (e.g., fried rice, flatbread) categories.
- **Description**: Annotators provide a description of each dish based on the content from its Wikipedia page, avoiding the use of the dish's name, origin, or any distinctive keywords that uniquely identify the dish.
- **Cuisine**: The dish's origin cuisine and any cuisines with which it is strongly associated.
- **Geographic Distribution**: This includes the dish's associated countries, area (city or region), and broader continental region.
### Quality Assurance
Before beginning the quality assurance process, we first identify common issues that arise during annotation and develop automated rules to detect easily identifiable annotation errors, such as incorrect string formatting. Annotators are then asked to correct these errors. To further ensure data quality and validity, we conduct several rounds of quality assurance:
1. **Image Quality**: We remove instances where images are blurry, dark, or contain distracting elements such as people or other dishes. We also verify image licenses by cross-referencing them with information on Wikimedia Commons.
2. **Categorization and Descriptions**: We refine dish categorization and descriptions, ensuring consistency in category assignments and keeping descriptions free from "information breaches" (e.g., excluding regional details from the description).
3. **Cuisine Names and Geographic Accuracy**: We standardize cuisine names and meticulously review all country and area information for accuracy.
This comprehensive approach guarantees the integrity and reliability of our dataset.
### Data Compilation
In this phase, we verify the overall quality check done by annotators, identifying any potential inconsistencies missed during quality assurance. We then compile the dataset by collecting the metadata into a single file.
## VQA Generation
In this phase, we generate VQA data by sampling from **WC-KB**. A VQA data entry comprises a visual image, question text, and answer text. This process involves four stages:
1. Conducting a similarity search for dish names
2. Constructing questions and contexts
3. Translating these elements into multiple languages
4. Generating the VQA triplets.
### Dish Names Similarity Search
To identify similar dishes in our dataset, we follow the approach from Winata et al. (2024) to employ a multilingual model E5$_\text{LARGE}$ Instruct (Wang et al. 2024) for computing text embeddings. Formally, given a dish $x$ with name $x_{\text{name}}$ and text description $x_{\text{desc}}$, we use a multilingual model $\theta$ to compute the embedding vector $v_x = \theta(\{x_\text{name};x_\text{desc}\})$, and then apply cosine similarity to compute a score $s = \text{similarity}(v_i, v_j)$ between dish $i$ and dish $j$. For each dish, we consider the top-$k$ most similar dishes to generate distractors in the multiple choice questions.
### Question and Context Construction
Dish name prediction (Task 1) is divided into three question variations depending on the context:
1. **No-context question**: Simply asks for the name of the dish without any provided context.
2. **Contextualized question**: Provides additional information related to cuisine or location.
3. **Adversarial contextualized question**: Similar to contextualized questions, but may include misleading location information to assess the model's robustness to irrelevant details.
For regional cuisine prediction (Task 2), only a basic question without any provided context is available.
### Multiple Language Translation
#### Question and Context
All questions and contexts are initially collected in English and are then carefully translated by native speakers into 30 language varieties: 23 different languages, with 7 languages having two varieties. Translators prioritize naturalness, followed by diversity in translations when duplication occurs.
#### Food Name Alias
Using Wikipedia pages as our primary source, we verify if the English page has translations available in other languages. This enables us to extract dish names in multiple languages and compile them as translations for each dish. We utilize both the Wikipedia page titles in various languages and alias text from the English page. These translations enhance cultural relevance and accuracy for multilingual prompts. When translation is unavailable, we use the English name as the default.
#### Locations and Cuisines
With over 400 unique locations, including countries, cities, and areas, we first translate the English locations into other languages using GPT-4, followed by proofreading by native speakers. Regional cuisine names (the adjective form of the location in English) are translated in the same manner.
#### Morphological Inflections
In languages with rich inflectional morphology (e.g., Czech or Spanish), words are modified to express different grammatical categories (e.g., number, gender, or case). We provide a framework for human translators to use natural inflections in the prompt template while keeping the inflections as few as possible.
### Generating VQA Triplets
To ensure no overlap between training and testing subsets, we split the dishes and multilingual questions into two subsets. For each subset, we randomly sample dish and question pairs. We use the dish entry in our KB dataset to select the image and inject the location into the context, if applicable. Answer candidates for multiple-choice questions are picked using the similarity search process. We repeat this process until the desired number of training or test samples is reached, discarding duplicates.
## Ethical Considerations
Our research focuses on evaluating VLMs within the context of multilingual and multicultural VQA, a field that holds significant implications for diverse multilingual communities. We are committed to conducting our data collection and evaluations with the highest standards of transparency and fairness. To achieve this, we have adopted a crowd-sourcing approach for the annotation process, inviting volunteers to contribute and become co-authors if they provide significant contributions. We follow the guidelines from ACL for authorship eligibility as shown in https://www.aclweb.org/adminwiki/index.php/Authorship_Changes_Policy_for_ACL_Conference_Papers. In line with our commitment to openness and collaboration, we will release our dataset under an open-source license, CC-BY-SA 4.0.
## Contact
E-mail: [Genta Indra Winata](genta.winata@capitalone.com) and [Frederikus Hudi](frederikus.hudi.fe7@is.naist.jp)
## Citation
If you find this dataset useful, please cite the following works
```bibtex
@article{winata2024worldcuisines,
title={WorldCuisines: A Massive-Scale Benchmark for Multilingual and Multicultural Visual Question Answering on Global Cuisines},
author={Winata, Genta Indra and Hudi, Frederikus and Irawan, Patrick Amadeus and Anugraha, David and Putri, Rifki Afina and Wang, Yutong and Nohejl, Adam and Prathama, Ubaidillah Ariq and Ousidhoum, Nedjma and Amriani, Afifa and others},
journal={arXiv preprint arXiv:2410.12705},
year={2024}
}
``` |