File size: 4,445 Bytes
7de4171
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7a7cfe2
 
 
 
7de4171
 
 
 
7a7cfe2
7de4171
0a360b5
 
 
 
 
 
55e3c9b
0a360b5
191e365
 
da8fd51
0a360b5
da8fd51
 
 
 
 
 
 
 
 
 
 
 
 
0a360b5
 
55e3c9b
 
 
 
 
 
 
da8fd51
0a360b5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bbe5578
0a360b5
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
---
dataset_info:
  features:
  - name: id
    dtype: int64
  - name: multimodal_question
    dtype: string
  - name: answer
    dtype: string
  - name: rationale
    dtype: string
  - name: text_only_question
    dtype: string
  - name: image_source
    dtype: string
  - name: evidence
    dtype: string
  - name: resolution
    dtype: string
  - name: proportion_of_roi
    dtype: string
  - name: category
    dtype: string
  - name: text_in_image
    dtype: string
  - name: rationale_granularity
    dtype: string
  - name: image
    dtype: image
  - name: cropped_image
    dtype: image
  splits:
  - name: train
    num_bytes: 157153160.0
    num_examples: 129
  download_size: 157133331
  dataset_size: 157153160.0
configs:
- config_name: default
  data_files:
  - split: train
    path: hard_data/train-*
---
# VisualSimpleQA

## Introduction
VisualSimpleQA is a multimodal fact-seeking benchmark with two key features. First, it enables streamlined and decoupled evaluation of LVLMs in visual and linguistic modalities. Second, it incorporates well-defined difficulty criteria to guide human annotation and facilitates the extraction of a challenging subset, VisualSimpleQA-hard.
Experiments on 15 LVLMs show that even state-of-the-art models such as GPT-4o achieve merely 60%+ correctness in multimodal fact-seeking QA on VisualSimpleQA and 30%+ on VisualSimpleQA-hard.
Furthermore, the decoupled evaluation based on this benchmark across different models highlights substantial opportunities for improvement in both visual and linguistic modules.
The dataset reviewer above illustrates 129 samples from VisualSimpleQA-hard.

`arXiv:` [https://arxiv.org/pdf/2503.06492](https://arxiv.org/pdf/2503.06492)

**Data Example:**
```
  {'id': 369,
   'multimodal_question': 'Which institution did the creator of this cartoon duck donate her natural science-related paintings to?',
   'answer': 'The Armitt Museum, Gallery, Library', 'rationale': 'Jemima Puddle-Duck',
   'text_only_question': 'Which institution did the creator of Jemima Puddle-Duck donate her natural science-related paintings to?',
   'image_source': 'https://www.gutenberg.org/files/14814/14814-h/images/15-tb.jpg',
   'evidence': 'https://www.armitt.com/beatrix-potter-exhibition/\nhttps://en.wikipedia.org/wiki/Beatrix_Potter',
   'resolution': '400x360',
   'proportion_of_roi': '0.2232',
   'category': 'research and education',
   'text_in_image': 'absence',
   'rationale_granularity': 'fine-grained',
   'image': <PIL.PngImagePlugin.PngImageFile image mode=RGB size=400x360 at 0x7FE82C270D70>,
   'cropped_image': <PIL.PngImagePlugin.PngImageFile image mode=RGB size=164x196 at 0x7FE82C329550>}
```

## File Structure

`data/`
This directory contains all 500 samples of VisualSimpleQA stored in parquet files.

`hard_data/`
This directory contains 129 VisualSimpleQA-hard samples stored in a parquet file. These samples are selected based on well-defined criteria to ensure they represent more challenging cases from VisualSimpleQA.

## Disclaimer

This dataset contains images collected from various sources. The authors do NOT claim ownership or copyright over the images. The images may be subject to third-party rights, and users are solely responsible for verifying the legal status of any content before use.

- Intended Use: The images are provided for ​non-commercial research purposes only.

- Redistribution Prohibition: You may NOT redistribute or modify the images without permission from original rights holders.

- Reporting Violations: If you encounter any sample potentially breaching copyright or licensing rules, contact us at [email protected]. Verified violations will be removed promptly.

The authors disclaim all liability for copyright infringement or misuse arising from the use of this dataset. Users assume full legal responsibility for their actions.

## License

- Text Data: Licensed under [CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/)  
- Images: Subject to custom terms (see Disclaimer above).

## Citation

**BibTeX:**
```bibtex
@article{wang2025visualsimpleqa,
  title={VisualSimpleQA: A Benchmark for Decoupled Evaluation of Large Vision-Language Models in Fact-Seeking Question Answering},
  author={Yanling Wang and Yihan Zhao and Xiaodong Chen and Shasha Guo and Lixin Liu and Haoyang Li and Yong Xiao and Jing Zhang and Qi Li and Ke Xu},
  journal={arXiv preprint arXiv: 2503.06492},
  year={2025}
}
```