Datasets:

Modalities:
Image
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
File size: 4,741 Bytes
9b870b2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
---
dataset_info:
  features:
  - name: image
    dtype: image
  - name: source
    dtype: string
  - name: question
    dtype: string
  - name: answer
    dtype: string
  - name: img_id
    dtype: string
  splits:
  - name: raw
    num_bytes: 15176464880.875
    num_examples: 58849
  download_size: 1547864596
  dataset_size: 15176464880.875
configs:
- config_name: default
  data_files:
  - split: raw
    path: data/*.parquet
license: cc-by-nc-4.0
task_categories:
- visual-question-answering
language:
- en
---

The Kvasir-VQA dataset is an extended dataset derived from the HyperKvasir and Kvasir-Instrument datasets, augmented with question-and-answer annotations. This dataset is designed to facilitate advanced machine learning tasks in gastrointestinal (GI) diagnostics, including image captioning, Visual Question Answering (VQA) and text-based generation of synthetic medical images

**Homepage**: [https://datasets.simula.no/kvasir-vqa](https://datasets.simula.no/kvasir-vqa)

# Usage
You can use the Kvasir-VQA dataset directly from HuggingFace Dataset Hub.

🔥 See [Jupyter Notebook Demo](https://huggingface.co/datasets/SushantGautam/SimulaMet-HOST/Kvasir-VQA/blob/main/explore_Kvasir-VQA.ipynb). You can open the notebook on Google Colab.

```
from datasets import load_dataset
ds = load_dataset("SimulaMet-HOST/Kvasir-VQA")
```

## Downloading Dataset as an Image foler and CSV Metadata

```
d_path ="./" #existing folder where you want to save images and metadata.csv

df = ds['train'].select_columns(['source', 'question', 'answer', 'img_id']).to_pandas()
df.to_csv(f"{d_path}/metadata.csv", index=False)

import os
os.makedirs(f"{d_path}/images", exist_ok=True)

for i, row in df.groupby('img_id').nth(0).iterrows(): # for images
  image = ds['train'][i]['image'].save(f"{d_path}/images/{row['img_id']}.jpg")
```

The total image size is around 1.5 GB. The CSV file will have 58,849 rows.


### Key Features

- **Total Images**: 6,500 annotated images
- **Annotations**: Includes question-and-answer pairs for each image
- **Question Types**: Yes/No, single-choice, multiple-choice, color-related, location-related, numerical count
- **Applications**: Image captioning, VQA, synthetic medical image generation, object detection, etc


## Dataset Details

### Image Categories

The dataset includes images from various GI tract conditions and medical instruments used in GI procedures:

| Image Category      | Number of Samples | Source Dataset      |
|---------------------|-------------------|---------------------|
| Normal              | 2500              | HyperKvasir         |
| Polyps              | 1000              | HyperKvasir         |
| Esophagitis         | 1000              | HyperKvasir         |
| Ulcerative Colitis  | 1000              | HyperKvasir         |
| Instrument          | 1000              | Kvasir-Instrument   |
| **TOTAL**           | **6500**          |                     |


### Annotation Process

Annotations were developed with input from medical professionals and include six types of questions:

- **Yes/No Questions**
- **Single-Choice Questions**
- **Multiple-Choice Questions**
- **Color-Related Questions**
- **Location-Related Questions**
- **Numerical Count Questions**

Annotations cover a range of GI aspects, including findings, abnormalities, anatomical landmarks, and medical instruments.


When using the Kvasir-VQA dataset, you should include the following information to ensure compliance with the dataset's usage terms, particularly when citing the dataset in documents or papers:

## Terms of Use

The data is released fully open for research and educational purposes under the [Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) license](https://creativecommons.org/licenses/by-nc/4.0/). The use of the dataset for purposes such as competitions and commercial purposes needs prior written permission. In all documents and papers that use or refer to the dataset or report experimental results based on the Kvasir-VQA, a reference to the related article needs to be added: 

```
@inproceedings{gautam2024kvasirvqa,
  title={Kvasir-VQA: A Text-Image Pair GI Tract Dataset},
  author={Gautam, Sushant and Storås, Andrea and Midoglu, Cise and Hicks, Steven A. and Thambawita, Vajira and Halvorsen, Pål and Riegler, Michael A.},
  booktitle={Proceedings of the First International Workshop on Vision-Language Models for Biomedical Applications (VLM4Bio '24)},
  year={2024},
  location={Melbourne, VIC, Australia},
  pages={10 pages},
  publisher={ACM},
  doi={10.1145/3689096.3689458}
}
```

## Contact
Please contact michael@simula.no, vajira@simula.no, steven@simula.no or paalh@simula.no for any questions regarding the dataset.