Datasets:

Modalities:
Audio
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas
License:
MMAU-Pro / README.md
sonalkum's picture
Update README.md
b8516c2 verified
---
license: cc-by-nc-4.0
---
# MMAU-Pro: A Challenging and Comprehensive Benchmark for Audio General Intelligence
[![Paper](https://img.shields.io/badge/arxiv-%20PDF-red)](https://www.arxiv.org/pdf/2508.13992) [![Audios](https://img.shields.io/badge/🔈%20-Audios-blue)](https://huggingface.co/datasets/gamma-lab-umd/MMAU-Pro/blob/main/data.zip)
[MMAU-Pro](https://arxiv.org/abs/2508.13992) is the most comprehensive benchmark to date for evaluating **audio intelligence in multimodal models**. It spans speech, environmental sounds, music, and their combinations—covering **49 distinct perceptual and reasoning skills**.
The dataset contains **5,305 expert-annotated question–answer pairs**, with audios sourced directly *from the wild*. It introduces several novel challenges overlooked by prior benchmarks, including:
- Long-form audio understanding (up to 10 minutes)
- Multi-audio reasoning
- Spatial audio perception
- Multicultural music reasoning
- Voice-based STEM and world-knowledge QA
- Instruction-following with verifiable constraints
- Open-ended QA in addition to MCQs
---
🚀 Usage
You can load the dataset via Hugging Face datasets:
```
from datasets import load_dataset
ds = load_dataset("gamma-lab-umd/MMAU-Pro")
```
For evaluation, we provide:
- MCQ scoring via embedding similarity (NV-Embed-v2)
- Open-ended QA with LLM-as-a-judge
- Regex based string matching for Instruction Following
---
🧪 Baselines & Model Performance
We benchmarked 22 leading models on MMAU-Pro.
- Gemini 2.5 Flash (closed-source): 59.2% avg. accuracy
- Audio Flamingo 3 (open-source): 51.7%
- Qwen2.5-Omni-7B: 52.2%
- Humans: ~78%
See full results in the paper.
---
🌍 Multicultural Music Coverage
MMAU-Pro includes music from 8 diverse regions:
• Western, Chinese, Indian, European, African, Latin American, Middle Eastern, Other Asian
This reveals clear biases: models perform well on Western/Chinese but poorly on Indian/Latin American music.
---
📥 Download
- Dataset: [HF](https://huggingface.co/datasets/gamma-lab-umd/MMAU-Pro)
- Paper: [MMAU-Pro](https://arxiv.org/abs/2508.13992)
- Website: [Official Page](https://sonalkum.github.io/mmau-pro/)
- Github: [Git](https://github.com/sonalkum/MMAUPro)
---
🧩 Evaluation
The evaluation code is designed to take in the complete `test.parquet` with predictions in the column `model_ouput`.
```
python evaluate_mmau_pro_comprehensive.py test.parquet --model_output_column model_output
```
---
✍️ Citation
If you use MMAU-Pro, please cite:
```bibtex
@article{kumar2025mmau,
title={MMAU-Pro: A Challenging and Comprehensive Benchmark for Holistic Evaluation of Audio General Intelligence},
author={Kumar, Sonal and Sedl{\'a}{\v{c}}ek, {\v{S}}imon and Lokegaonkar, Vaibhavi and L{\'o}pez, Fernando and Yu, Wenyi and Anand, Nishit and Ryu, Hyeonggon and Chen, Lichang and Pli{\v{c}}ka, Maxim and Hlav{\'a}{\v{c}}ek, Miroslav and others},
journal={arXiv preprint arXiv:2508.13992},
year={2025}
}
```
---
🙏 Acknowledgments
Some work was carried out at JSALT 2025.