MultihopSpatial: Multi-hop Compositional Spatial Reasoning Benchmark for Vision-Language Model

* Equal contribution

1
2
3
4

TL;DR

Are Vision-Language Models robust at multi-hop compositional spatial reasoning?

Not yet. In real-world scenarios, human instructions to embodied agents are far from simple — they require chaining multiple spatial inferences such as perspective-taking, attribute filtering, and relation comparison. However, existing benchmarks only test single-hop spatial relations (e.g., "Is X to the right of Y?"), which cannot capture this compositional complexity. We introduce MultihopSpatial, a benchmark featuring 1- to 3-hop compositional spatial queries paired with visual grounding evaluation. Our extensive evaluation of 37 VLMs reveals that multi-hop spatial reasoning remains a formidable challenge even for frontier models, and that models can often select the correct MCQ answer without genuinely locating the target — a critical blind spot we expose through our grounded metric, Acc@50IoU.

Comparison of existing benchmarks and MultihopSpatial

Figure: (Left) Existing benchmarks rely on single-hop standard MCQs without spatial localization. MultihopSpatial requires multi-hop compositional reasoning with grounded evaluation. (Right) Frontier VLMs show high MCQ accuracy (solid bars) but plummet under Acc@50IoU (striped bars), revealing a severe spatial grounding blind spot.

Multi-hop Compositional Reasoning

Real-world instructions require chaining multiple spatial inferences. MultihopSpatial introduces 1- to 3-hop queries composing attribute, position, and relation — going far beyond the single-step "left/right" questions of existing benchmarks.

Motivated by Embodied AI

Instructions to embodied agents (e.g., humanoid robots, VLAs) are inherently complex. An agent must adopt a perspective, filter candidates, and compare relations before it can act — exactly the multi-hop reasoning our benchmark evaluates.

Grounded Evaluation

Embodied agents need not only correct answers but also precise visual grounding. We propose Acc@50IoU, which requires both a correct MCQ answer and an accurate bounding box, exposing models that guess without truly locating the target.

Training Corpus for Spatial Intelligence

MultihopSpatial-Train (6,791 samples) enables RL post-training that enhances intrinsic spatial reasoning across 5 benchmarks and translates to improved downstream manipulation performance in VLA tasks.

Research Motivation

Why multi-hop spatial reasoning matters for embodied AI

Multi-hop spatial reasoning for embodied agents

Figure: A multi-hop spatial reasoning example for an embodied agent in a real-world scenario.

The development of embodied agents — particularly Vision-Language-Action (VLA) models — fundamentally relies on VLMs for spatial reasoning to interact with the physical world. In complex real-world environments, an instruction like "Could you move the round cup on my right — the one furthest away?" requires an agent to adopt an ego-centric perspective, isolate the target position, filter by attributes, and compare spatial relations.

This internal reasoning mirrors a diagnostic multi-hop multiple-choice question coupled with precise bounding box prediction. An agent can only successfully navigate and manipulate an object if it accurately answers the query and visually grounds the target. However, existing spatial reasoning benchmarks focus almost exclusively on single-hop queries (e.g., "Is X to the right of Y?"), failing to capture the compositional, multi-step reasoning essential for real-world embodied scenarios.

MultihopSpatial

A comprehensive benchmark for multi-hop compositional spatial reasoning with visual grounding

4,500
Benchmark QA Pairs
6,791
Training Samples
3,563
Curated Images
1–3
Reasoning Hops
90%
Inter-Rater Agreement (Krippendorff's α)

MultihopSpatial is a 4,500-sample benchmark of manually annotated VQA pairs focusing on multi-hop, compositional spatial queries common in real-world scenarios. Each example provides ground-truth bounding boxes to jointly evaluate reasoning and spatial localization, which is crucial for embodied/VLA settings (e.g., grasping and action). This grounded design improves interpretability and eliminates the evaluation blind spot of random guessing.

We curate 3,563 spatially complex images from COCO and PACO-Ego4D, ensuring diverse coverage of everyday indoor/outdoor scenes and ego/exo perspectives. Upon these images, we construct 4,500 MCQs, perfectly balanced across 1- to 3-hop reasoning levels (1,500 per hop) and viewpoints (750 ego-centric and 750 exo-centric per hop).

To eliminate the reliability concerns and hallucinations inherent in AI-generated data, all QA pairs and bounding boxes were strictly annotated by ten trained human experts. Each sample underwent a rigorous multi-stage verification with three rounds of independent cross-checking, achieving high inter-annotator agreement (Krippendorff's α = 0.90). Furthermore, we introduce MultihopSpatial-Train, an auxiliary large-scale corpus of 6,791 grounded VQA samples designed to support post-training of VLMs for spatial intelligence.

Compositional structure and category definitions

Figure: Compositional structure and category definitions in MultihopSpatial.

Spatial Reasoning Categories

We define three fundamental spatial reasoning categories: Attribute (att), Position (pos), and Relation (rel). Our benchmark systematically composes these categories into multi-hop questions (e.g., 2- and 3-hop) that demand sequential intermediate inferences — identifying candidates, comparing relations, and resolving ambiguities. By increasing the hop count, we represent longer reasoning chains and escalating difficulty, enabling fine-grained diagnosis of model performance across different levels of reasoning complexity.

1-Hop: Single-step questions targeting one spatial category (pos or rel). We exclude att as a standalone category since attributes are primarily perceptual unless composed with spatial metrics. While 1-hop spatial reasoning has been explored in prior works, we deliberately include it as a controlled baseline for depth-wise comparisons against multi-hop compositions.

2-Hop: Questions combining two categories (att+pos, att+rel, or pos+rel). These queries follow a two-stage structure: (i) restricting the candidate set using one category, and (ii) identifying the target by applying the other. Both constraints must be jointly satisfied, regardless of the specific inference order.

3-Hop: Questions incorporating all three categories (att+pos+rel) in a single query. An att cue narrows candidates, after which the model reasons over pos and rel to identify the target (e.g., selecting the rightmost object that is farthest/closest). This structure mirrors how humans refer to objects in cluttered scenes, testing the disambiguation needed for embodied task execution.

Example MultihopSpatial questions

Figure: Example MultihopSpatial questions for 2-hop and 3-hop reasoning under ego-centric (top) and exo-centric (bottom) views.

Benchmark Results

Evaluating 37 VLMs across 5 categories on MultihopSpatial

Evaluation Metrics

MCQ Accuracy

Percentage of correct multiple-choice predictions. Standard but does not verify spatial localization.

Acc@50IoU

Our primary grounded metric. Correct only if the answer matches and predicted bounding box has IoU ≥ 0.5 with ground truth.

Avg. IoU

Computed over MCQ-correct samples only, isolating grounding capability from reasoning errors.

Main Results

Overall Performance: Gemini-3-Pro achieves the highest MCQ accuracy (64.7%) and Acc@50IoU (40.6%), while Qwen3-VL-32B-Thinking leads the open-weight category. Top performance in answer selection does not guarantee precise localization.

Metric-dependent Rankings: Claude-Opus-4.5 drops from 7th (MCQ) to 29th (Acc@50IoU), while the smaller Qwen3-VL-4B rises from 25th to 10th due to robust grounding — proving MCQ alone is misleading.

Benchmark Difficulty: The best model peaks at just 40.6% Acc@50IoU. Under 3-hop ego-centric conditions, only 3 of 37 models exceed the 25% random MCQ baseline. GPT-5.2-Thinking achieves just 8.5% and Claude-Sonnet-4.5-Thinking just 1.9%.

Model Overall 3Hop-Ego 3Hop-Exo 2Hop-Ego 2Hop-Exo 1Hop-Ego 1Hop-Exo
Acc.Acc@50IoU Acc.Acc@50 Acc.Acc@50 Acc.Acc@50 Acc.Acc@50 Acc.Acc@50 Acc.Acc@50
Proprietary Models — Instant
Claude-Opus-4.5 45.13.213.3 25.72.0 48.53.6 33.72.0 58.14.8 43.23.5 61.13.1
Claude-Sonnet-4.5 20.90.54.2 6.40.4 18.50.9 12.30.0 34.71.6 13.50.0 40.00.1
GPT-5.2 19.32.011.8 5.30.7 18.04.9 10.80.1 30.45.9 8.10.0 43.30.4
Open-weight Models — Instant
GLM-4.6V 43.235.269.5 15.912.3 46.739.3 22.718.4 61.653.2 32.424.3 80.163.7
Molmo2-8B 41.80.38.8 15.90.3 44.40.4 21.60.3 60.40.1 32.40.4 76.40.3
Qwen3-VL-235B 41.334.871.1 14.812.3 42.937.6 21.918.5 58.552.7 30.723.5 79.264.4
Qwen3-VL-32B 40.933.469.6 13.99.7 43.936.4 21.917.2 59.253.1 30.422.3 76.461.6
InternVL-3.5-38B 40.89.728.7 17.53.2 44.512.3 24.34.5 56.824.5 31.93.3 69.610.4
InternVL-3.5-14B 39.77.926.2 17.12.5 44.510.1 22.02.8 56.114.4 30.43.3 68.014.0
InternVL-3.5-4B 39.710.330.0 15.93.1 45.213.5 23.94.5 54.918.7 31.93.6 66.318.4
InternVL-3.5-8B 38.39.430.0 14.72.9 41.110.7 23.34.0 50.117.2 30.54.5 70.017.3
Qwen3-VL-8B 38.031.369.5 12.38.8 42.136.1 18.815.2 55.349.1 26.720.8 72.858.0
Qwen3-VL-4B 37.831.069.9 15.510.9 40.133.7 20.916.0 53.546.9 26.721.2 70.357.2
Gemma-3-IT-27B 33.10.45.4 18.10.1 30.80.5 22.00.1 45.71.1 28.10.3 53.90.3
Gemma-3-IT-12B 29.80.45.9 16.90.3 31.60.5 22.10.4 40.80.5 22.90.1 44.50.5
Gemma-3-IT-4B 28.40.23.0 22.10.0 27.20.3 22.50.1 36.80.3 27.10.3 34.70.1
Proprietary Models — Reasoning
Gemini-3-Pro* 64.740.655.0 39.718.8 71.145.3 36.820.5 81.255.5 71.141.1 88.462.3
GPT-5.2-Thinking 57.911.529.0 36.18.5 55.710.4 49.77.6 63.618.0 65.612.5 76.411.7
Gemini-3-Flash* 57.240.261.2 6.94.3 61.246.9 42.325.3 80.063.7 66.038.9 86.862.1
Claude-Opus-4.5-Thinking 47.04.716.7 25.53.5 49.74.7 35.23.1 60.08.8 45.15.2 66.52.9
Claude-Sonnet-4.5-Thinking 32.24.319.2 14.71.9 29.93.6 22.12.3 45.78.1 31.33.9 49.36.1
Open-weight Models — Reasoning
Qwen3-VL-32B-Thinking 46.837.467.2 19.212.9 57.547.1 24.318.1 70.160.0 30.423.1 79.663.2
Qwen3-VL-235B-Thinking 45.136.367.8 17.612.7 51.242.3 24.819.3 67.658.7 31.222.8 78.161.9
Qwen3-VL-4B-Thinking 42.628.758.3 20.49.2 48.134.8 22.312.0 63.950.7 30.816.3 70.449.2
InternVL-3.5-38B-Thinking 42.127.457.0 19.510.9 43.332.5 24.715.3 56.839.2 34.820.7 73.645.9
GLM-4.6V-Thinking 42.034.770.1 14.110.7 46.340.0 19.515.3 63.156.1 31.223.3 77.662.7
Qwen3-VL-8B-Thinking 41.729.560.1 18.59.2 47.936.4 21.311.9 63.351.6 28.516.5 70.551.2
InternVL-3.5-8B-Thinking 40.65.222.1 20.51.3 39.96.7 24.11.9 57.310.4 30.93.5 70.57.6
InternVL-3.5-4B-Thinking 40.64.721.4 18.31.2 41.24.8 24.41.7 58.59.6 32.81.5 68.19.3
InternVL-3.5-14B-Thinking 38.211.134.7 14.14.7 42.813.6 21.15.6 55.320.5 27.66.4 68.015.9
Specialized Spatial Reasoning Models
SenseNova-InternVL3-8B 42.317.338.8 20.49.1 45.219.7 25.29.2 55.527.2 34.511.2 73.227.2
Cosmos-Reason2-8B 37.827.961.4 15.210.5 40.731.9 19.513.5 54.543.5 26.517.1 70.151.1
VST-7B-RL 36.00.01.5 16.70.0 34.10.1 23.90.0 48.10.0 24.50.0 68.80.0
SpaceQwen3-VL-2B 33.610.131.5 18.54.0 32.59.9 22.84.0 47.222.9 26.14.4 54.315.2
SpaceOm 32.30.32.6 15.30.5 37.90.1 19.60.1 47.90.4 20.50.3 52.80.4
SpatialReasoner 31.78.729.9 18.06.8 34.010.3 19.64.3 46.013.7 21.35.7 51.511.2
SpaceThinker-3B 31.14.016.6 15.92.9 36.33.7 19.22.9 44.56.5 20.83.3 50.04.3

Table 1: Benchmark results across different hop counts and Ego/Exo perspectives. Green cells indicate the best performance within each model group. *Gemini-3-Pro & -Flash operate with thinking mode enabled by default and do not provide an option to fully disable it. We therefore classify them as reasoning models.

In-depth Analysis

Eight key insights from evaluating 37 VLMs on MultihopSpatial

1

Performance Degradation across Reasoning Hops

Evaluating the impact of reasoning complexity across all 37 models reveals a consistent performance degradation as the number of hops increases, confirming that compositional spatial reasoning remains a fundamental challenge for current VLMs. Both MCQ accuracy and Acc@50IoU exhibit steep declines from 1-hop to 3-hop.

Crucially, this degradation is exacerbated under ego-centric evaluation, which requires additional perspective transformation. The widening performance gap between ego-centric and exo-centric views at higher hops suggests that perspective-taking compounds with multi-step reasoning, creating a multiplicative rather than additive difficulty.

Average Performance by Hop Count

Average performance by hop count. Performance degrades consistently as reasoning hops increase.

2

Ego vs. Exo: Perspective-Taking as a Compounding Bottleneck

Beyond a simple performance drop, ego-centric evaluation fundamentally alters capability visibility. Under exo-centric conditions, our Acc@50IoU metric clearly distinguishes grounding-capable models (e.g., Qwen3-VL) from those lacking native localization (e.g., InternVL-3.5) despite their similar MCQ accuracies.

Conversely, ego-centric evaluation acts as an evaluation blind spot: it suppresses even strong grounding models to a 20–25% floor, completely masking these capability gaps. This compression highlights the necessity of evaluating perspectives jointly and reinforces Acc@50IoU as an essential metric to expose disparities invisible to conventional MCQ evaluation.

Scatter plots across perspectives

Model performance scatter plots. Ego-centric evaluation masks grounding differences visible in exo-centric conditions.

3

Instant vs. Reasoning: Diminishing Returns under Multi-Hop Pressure

While reasoning models consistently outperform instant models, this advantage diminishes under multi-hop pressure. The initial MCQ gain of up to +8 pp at 1-hop narrows considerably by 3-hop.

Critically, even with extended thinking, reasoning models plummet to sub-20% MCQ accuracy and sub-10% Acc@50IoU on 3-hop ego-centric tasks. This convergence toward the performance floor reveals that inference-time reasoning (e.g., chain-of-thought) yields diminishing returns as compositional steps accumulate. MultihopSpatial thus exposes a capability ceiling that test-time compute alone cannot resolve.

Reasoning vs. Instant Models

Reasoning vs. Instant models. The reasoning advantage shrinks with increasing hop count.

4

Open-weight vs. Proprietary Models

While proprietary models hold a slight MCQ advantage in ego-centric tasks—indicating stronger perspective-taking—open-weight models consistently dominate in Acc@50IoU. This reversal stems from a distinct asymmetry in visual grounding capabilities.

Proprietary models exhibit high variance: while the Gemini-3 series demonstrates robust localization, others (e.g., GPT-5.2, Claude) lack native grounding, significantly dragging down the group average. Conversely, open-weight families (e.g., Qwen3-VL, GLM-4.6V) maintain consistently high grounding performance, forming a dense cluster at 60%+ in exo-centric Acc@50IoU.

Open-weight vs. Proprietary models

Open-weight vs. Proprietary models across hop count and viewpoint.

5

Generalist vs. Specialist Models

Comparing specialized spatial reasoning models (SRMs) against comparable general-purpose VLMs (≤10B) reveals a counterintuitive trend: SRMs consistently underperform across all metrics and hop counts.

This shortfall primarily stems from the robust native grounding inherent in general models (e.g., Qwen3-VL), which most SRMs lack. These findings indicate that current SRMs, typically fine-tuned on single-step QA, fail to generalize to multi-hop scenarios, suggesting that general-purpose VLMs provide a stronger foundation for compositional spatial understanding.

Spatial Reasoning Models vs. General Models

Spatial Reasoning Models vs. General Models (≤10B).

6

Grounding Gap: Grounded vs. Ungrounded Accuracy

A striking disconnect exists between answer selection and spatial localization: most models fall far below the y=x line, indicating that high MCQ accuracy does not entail accurate grounding. Across all 37 models, the average ungrounded ratio reaches 59%, meaning more than half of correctly answered questions lack proper spatial localization.

This ratio varies dramatically by category: proprietary instant models exhibit 93% ungrounded accuracy, while open-weight reasoning models achieve the lowest at 43%. At the extreme, models such as Gemma-3-IT, Molmo2, and Claude-Sonnet-4.5 exceed 98% ungrounded, effectively answering through shortcuts without any spatial understanding.

Grounded vs Ungrounded accuracy per model

Grounded (Acc@50IoU) vs. Ungrounded accuracy per model. Hatched regions indicate ungrounded responses.

Grounding Gap

Grounding Gap: MCQ Acc. vs. Acc@50IoU.

Ungrounded Ratio by Category

Ungrounded Ratio by Model Category.

7

The Limits of LLM Scaling in Spatial Reasoning

Analyzing model scale across three open-weight families reveals that scaling the language backbone alone yields limited gains. While MCQ accuracy shows modest, saturating improvements, Acc@50IoU largely plateaus (e.g., Qwen3-VL) or remains near zero (e.g., Gemma-3-IT).

A notable exception is InternVL-3.5 at 38B, which exhibits a sharp grounding jump from 5.2% to 27.4%. Crucially, this leap coincides with a significant upgrade in its vision encoder (from 300M to 6B), whereas models utilizing fixed, small vision encoders (e.g., Qwen3-VL’s 400M) saturate early. This demonstrates that multi-hop spatial reasoning depends critically on scaling the capacity of visual spatial representations, not solely the language component.

Effect of model scale

Effect of model scale on MCQ Accuracy (top) and Acc@50IoU (bottom) across overall and per-hop settings.

8

Error Analysis on Tag Compositions

Analyzing error rates across spatial tag combinations reveals that while reasoning models consistently outperform instant models, multi-tag compositions remain a critical bottleneck. Specifically, the Position-Relation (P-R) combination incurs significantly higher errors than single-tag settings, highlighting the severe difficulty of jointly handling positional localization and relational comparison.

Even specialized spatial reasoning models, which achieve the lowest overall errors, exhibit non-trivial failures on these complex compositions. This underscores that true compositional spatial reasoning remains an unresolved challenge across the current VLM ecosystem.

Error rates by tag combination

Average error rates (%) by tag combination across model categories. A, R, and P denote Attribute, Relation, and Position.

MultihopSpatial-Train: Utility as a Training Corpus

Reinforcement learning post-training enhances both spatial reasoning and downstream VLA performance

Beyond serving as a static evaluation benchmark, we investigate MultihopSpatial as a training corpus. We post-train Qwen3-VL-4B-Instruct via Group Relative Policy Optimization (GRPO) using a composite reward function that jointly evaluates format adherence, answer correctness, and spatial localization accuracy:

R = Rformat + α · Rmcq + β · Rbbox

Format Reward: Binary reward (1 if output follows expected parsing format, 0 otherwise).

MCQ Reward: Discrete correctness signal (1 if predicted choice matches ground truth).

BBox Reward: Continuous GIoU-based localization score normalized to [0, 1], providing dense geometric feedback.

Spatial Reasoning Benchmarks

The GRPO-trained model achieves substantial improvements on both the in-domain MultihopSpatial benchmark and five out-of-domain spatial reasoning benchmarks, demonstrating that spatial reasoning capabilities generalize beyond the training distribution.

Model MultihopSpatial (In-domain) Out-of-Domain
Acc.Acc@50IoUavg. IoU BLINK3DSRBenchOmniSpatialVSI-BenchSpatialMQA
Qwen3-VL-4B-Instruct 37.831.069.9 82.556.142.762.839.6
w/ MultihopSpatial 62.953.872.6 85.356.343.963.241.1

Table 2: Impact of MultihopSpatial-Train on spatial reasoning across diverse benchmarks.

VLM as Backbone for Vision-Language-Action

When deployed as a VLM backbone using VLM4VLA, the MultihopSpatial-trained model consistently outperforms the baseline. On CALVIN, the average task completion score rises from 3.75 to 3.98, with the performance gap widening as the task chain lengthens (from +0.6% on Task-1 to +7.0% on Task-5). The model also achieves a +4.2% improvement on Libero (35.8% → 40.0%).

VLM backbone Task-1Task-2Task-3Task-4Task-5 CalvinLibero
Qwen3-VL-4B-Instruct 92.481.874.166.859.9 3.7535.8
w/ MultihopSpatial-Train 93.085.479.373.266.9 3.9840.0

Table 3: Vision-Language-Action evaluation on Calvin ABC-D and Libero. Each task measures success rate (%). Calvin denotes the average number of successfully completed tasks per sequence.

Qualitative Results

Illustrative failure cases revealing cascading errors in multi-hop spatial reasoning

Qualitative failure analysis

Figure: Qualitative failure analysis on a 3-hop ego-centric question. All three models recognized the position condition "in front of," but failed to incorporate it into their final predictions.

The query asks to identify the farthest square or rectangular object in front of the sitting person. However, all three models neglect the position condition, instead selecting the farthest rectangular object based solely on attribute and relation. Examining each model's rationale reveals that they explicitly acknowledge the "in front of" constraint during reasoning, yet consistently disregard it in the final answer — demonstrating a failure to maintain intermediate conditions throughout the chain. An error at any single hop inevitably propagates to an incorrect final prediction.

Additional Qualitative Examples

Browse through representative failure cases across 2-hop and 3-hop ego-centric and exo-centric questions:

Citation

@article{lee2026multihopspatial,
  title={MultihopSpatial: Multi-hop Compositional Spatial Reasoning Benchmark for Vision-Language Model},
  author={Lee, Youngwan and Jang, Soojin and Cho, Yoorhim and Lee, Seunghwan and Lee, Yong-Ju and Hwang, Sung Ju},
  journal={arXiv preprint arXiv:2603.18892},
  year={2026},
  url={https://arxiv.org/abs/2603.18892}
}