We present OmniSpatial, a benchmark for evaluating spatial reasoning abilities in vision-language models.
It covers four major categories and 50+ subtypes, totaling over 1.5K QA pairs. Experiments reveal significant
limitations in existing models and suggest directions for improvement.