Datasets:
hezheqi
commited on
Commit
•
b3b6556
1
Parent(s):
21a65cb
Update readme
Browse files
README.md
CHANGED
@@ -43,9 +43,11 @@ configs:
|
|
43 |
- "val/*.parquet"
|
44 |
---
|
45 |
# CMMU
|
46 |
-
[**📖 Paper**](https://arxiv.org/) | [**🤗 Dataset**](https://huggingface.co/datasets) | [**GitHub**](https://github.com/FlagOpen/CMMU)
|
47 |
|
48 |
-
This repo contains the evaluation code for the paper [**CMMU: A Benchmark for Chinese Multi-modal Multi-type Question Understanding and Reasoning**](https://arxiv.org/) .
|
|
|
|
|
49 |
|
50 |
## Introduction
|
51 |
CMMU is a novel multi-modal benchmark designed to evaluate domain-specific knowledge across seven foundational subjects: math, biology, physics, chemistry, geography, politics, and history. It comprises 3603 questions, incorporating text and images, drawn from a range of Chinese exams. Spanning primary to high school levels, CMMU offers a thorough evaluation of model capabilities across different educational stages.
|
@@ -74,7 +76,7 @@ We currently evaluated 10 models on CMMU. The results are shown in the following
|
|
74 |
@article{he2024cmmu,
|
75 |
title={CMMU: A Benchmark for Chinese Multi-modal Multi-type Question Understanding and Reasoning},
|
76 |
author={Zheqi He, Xinya Wu, Pengfei Zhou, Richeng Xuan, Guang Liu, Xi Yang, Qiannan Zhu and Hua Huang},
|
77 |
-
|
78 |
year={2024},
|
79 |
}
|
80 |
```
|
|
|
43 |
- "val/*.parquet"
|
44 |
---
|
45 |
# CMMU
|
46 |
+
[**📖 Paper**](https://arxiv.org/abs/2401.14011) | [**🤗 Dataset**](https://huggingface.co/datasets) | [**GitHub**](https://github.com/FlagOpen/CMMU)
|
47 |
|
48 |
+
This repo contains the evaluation code for the paper [**CMMU: A Benchmark for Chinese Multi-modal Multi-type Question Understanding and Reasoning**](https://arxiv.org/abs/2401.14011) .
|
49 |
+
|
50 |
+
We release the validation set of CMMU, you can download it from [here](https://huggingface.co/datasets/BAAI/CMMU). The test set will be hosted on the [flageval platform](https://flageval.baai.ac.cn/). Users can test by uploading their models.
|
51 |
|
52 |
## Introduction
|
53 |
CMMU is a novel multi-modal benchmark designed to evaluate domain-specific knowledge across seven foundational subjects: math, biology, physics, chemistry, geography, politics, and history. It comprises 3603 questions, incorporating text and images, drawn from a range of Chinese exams. Spanning primary to high school levels, CMMU offers a thorough evaluation of model capabilities across different educational stages.
|
|
|
76 |
@article{he2024cmmu,
|
77 |
title={CMMU: A Benchmark for Chinese Multi-modal Multi-type Question Understanding and Reasoning},
|
78 |
author={Zheqi He, Xinya Wu, Pengfei Zhou, Richeng Xuan, Guang Liu, Xi Yang, Qiannan Zhu and Hua Huang},
|
79 |
+
journal={arXiv preprint arXiv:2401.14011},
|
80 |
year={2024},
|
81 |
}
|
82 |
```
|