Datasets:
image
imagewidth (px) 640
640
|
---|
e
e
Spatial457: A Diagnostic Benchmark for 6D Spatial Reasoning of Large Multimodal Models
Xingrui Wang1, Wufei Ma1, Tiezheng Zhang1, Celso M de Melo2, Jieneng Chen1, Alan Yuille1
1 Johns Hopkins University 2 DEVCOM Army Research Laboratory
Project Page / Paper / 🤗 Huggingface / Code
🧠 Introduction
Spatial457 is a diagnostic benchmark designed to evaluate the 6D spatial reasoning capabilities of large multimodal models (LMMs). It systematically introduces four key capabilities—multi-object understanding, 2D and 3D localization, and 3D orientation—across five difficulty levels and seven question types, progressing from basic recognition to complex physical interaction.
📦 Download
You can access the full dataset and evaluation toolkit:
- Dataset: Hugging Face
- Code: GitHub Repository
- Paper: arXiv 2502.08636
📊 Benchmark
We benchmarked a wide range of state-of-the-art models—including GPT-4o, Gemini, Claude, and several open-source LMMs—on all subsets. Performance consistently drops as task difficulty increases. PO3D-VQA and humans remain most robust across all levels.
The table below summarizes model performance across 7 subsets:
📚 Citation
@inproceedings{wang2025spatial457,
title = {Spatial457: A Diagnostic Benchmark for 6D Spatial Reasoning of Large Multimodal Models},
author = {Wang, Xingrui and Ma, Wufei and Zhang, Tiezheng and de Melo, Celso M and Chen, Jieneng and Yuille, Alan},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2025},
url = {https://arxiv.org/abs/2502.08636}
}
- Downloads last month
- 129