Papers
Topics
Authors
Recent
Search
2000 character limit reached

HopChain: Multi-Hop Data Synthesis for Generalizable Vision-Language Reasoning

Published 17 Mar 2026 in cs.CV, cs.AI, and cs.CL | (2603.17024v2)

Abstract: Vision-LLMs (VLMs) show strong multimodal capabilities but still struggle with fine-grained vision-language reasoning. We find that long chain-of-thought (CoT) reasoning exposes diverse failure modes, including perception, reasoning, knowledge, and hallucination errors, which can compound across intermediate steps. However, most existing vision-language data used for reinforcement learning with verifiable rewards (RLVR) does not involve complex reasoning chains that rely on visual evidence throughout, leaving these weaknesses largely unexposed. We therefore propose HopChain, a scalable framework for synthesizing multi-hop vision-language reasoning data for RLVR training of VLMs. Each synthesized multi-hop query forms a logically dependent chain of instance-grounded hops, where earlier hops establish the instances, sets, or conditions needed for later hops, while the final answer remains a specific, unambiguous number suitable for verifiable rewards. We train Qwen3.5-35B-A3B and Qwen3.5-397B-A17B under two RLVR settings: the original data alone, and the original data plus HopChain's multi-hop data, and compare them across 24 benchmarks spanning STEM and Puzzle, General VQA, Text Recognition and Document Understanding, and Video Understanding. Although this multi-hop data is not synthesized for any specific benchmark, it improves 20 of 24 benchmarks on both models, indicating broad and generalizable gains. Consistently, replacing full chained queries with half-multi-hop or single-hop variants reduces the average score across five representative benchmarks from 70.4 to 66.7 and 64.3, respectively. Notably, multi-hop gains peak in long-CoT vision-language reasoning, exceeding 50 points in the ultra-long-CoT regime. These experiments establish HopChain as an effective, scalable framework for synthesizing multi-hop data that improves generalizable vision-language reasoning.

Summary

  • The paper introduces HopChain, a novel multi-hop framework that constructs logically dependent visual queries to counter compounding reasoning errors.
  • The methodology integrates explicit perception-level hops with systematic instance segmentation to ensure precise visual grounding across logical steps.
  • Empirical validation shows significant gains on 20 out of 24 benchmarks, demonstrating enhanced generalization in extended chain-of-thought tasks.

HopChain: Multi-Hop Data Synthesis for Generalizable Vision-Language Reasoning

Motivation and Problem Analysis

Despite improving VLMs using RLVR, fine-grained and robust vision-language reasoning remains a significant challenge, especially under long chain-of-thought (CoT) prompting. When forced to perform multi-step visuo-linguistic chains, models frequently exhibit compounded failure types—including fine-grained perception errors, mishandled reasoning, knowledge gaps, and hallucinated intermediate steps. Standard vision-language RLVR training data typically lack such complex, dependency-structured chains, thus failing to sufficiently expose or correct these systemic weaknesses. Figure 1

Figure 1: Overview of the HopChain paradigm, highlighting how synthesized, visually grounded multi-hop questions address long-CoT compounding error failure modes prevalent in standard datasets.

Analysis of model failures demonstrates that mistakes in early intermediate steps—such as misidentifying an object—cascade, corrupting all downstream reasoning and answer synthesis. Empirical annotation of errors shows a broad distribution spanning perception and reasoning as dominant categories, with knowledge and hallucination errors also significantly represented. Figure 2

Figure 2: Distribution of error types before and after inclusion of multi-hop data, illustrating the dominance of perception and reasoning failures and the broad effect of multi-hop augmentation.

Figure 3

Figure 3: Qualitative examples where models without multi-hop training exhibit unreliable perception in extended reasoning chains, with highlighted corrections from multi-hop-augmented models.

The HopChain Framework

HopChain presents a scalable, systematic pipeline for generating structurally rich, multi-hop vision-language RLVR queries. Each synthesized example forms a chain where each hop is logically tied to entities, sets, or predicates resolved in prior hops—enforcing repeated visual regrounding and preventing language-only shortcutting.

Query Structure Formalization

HopChain's tasks are structurally enforced along two orthogonal dimensions:

  • Perception-level hops: Alternation between single-object (attribute reading, text recognition, localization) and multi-object (counting, relational judgment, comparison) operations.
  • Instance-chain hops: Explicit dependency chains where identification of the next entity/set requires resolving all prior hops (e.g., A→B→CA \to B \to C), tightly coupling each query step to cumulative visual evidence.

Each synthesized query terminates in an unambiguous numerical answer compatible with RLVR verifiability, ensuring that correct final output requires persistent accuracy at every hop. Figure 4

Figure 4: Examples of multi-hop data synthesized by HopChain, showing explicit instance chains and perception-level transitions grounded in visual attributes across image regions.

Scalable Synthesis Pipeline

The HopChain pipeline comprises:

  1. Category Identification: Semantic parsing of image-level entities using a high-capacity VLM to enumerate target categories.
  2. Instance Segmentation: Localization of individual object instances via SAM3 to generate meaningful referents for subsequent hops.
  3. Multi-Hop Query Generation: Automated and constrained construction of logical chains using a strong VLM with explicit dependence on visual evidence, ensuring query answerability only from real image content and preventing auxiliary signal leakage.
  4. Human-in-the-loop Annotation & Calibration: Multiple annotators independently solve each candidate query-labeled image, retaining only consensus-verified cases, after which overly easy queries are filtered using weaker models to maintain a challenging, informative training set.

Empirical Validation

Broad-Benchmark Generalization

Augmenting standard RLVR data with HopChain's multi-hop dataset yields robust, benchmark-agnostic improvements. Neither Qwen3.5-35B-A3B nor Qwen3.5-397B-A17B are tuned to any specific benchmark, yet multi-hop augmentation yields improvement on 20 out of 24 benchmarks across STEM, puzzle, general VQA, document understanding, and video understanding domains.

Ablation confirms that preservation of full query chains is critical: truncation to only single-hop or late-hop variants reduces mean benchmark scores substantially (from 70.4 to 66.7 and 64.3). Gains are most emphatic in long-response (ultra-long-CoT) regimes, with improvements exceeding 50 accuracy points in the longest bin. Figure 5

Figure 5: Comparative benchmark results under single-hop, half-multi-hop, and full multi-hop training; showing consistent advantage for full multi-hop across diverse tasks.

Figure 6

Figure 6: Accuracy as a function of response length, demonstrating that improvements from multi-hop data are especially pronounced for longer, more compound reasoning chains.

Difficulty Spectrum and Error-Mode Correction

Analysis reveals the synthesized multi-hop data exhibits substantial difficulty range, as evidenced by the distribution of partially, sparsely, and fully solved queries by both mid- and large-scale models. Figure 7

Figure 7: Success-rate distribution of multi-hop data, confirming a broad difficulty profile appropriate for driving model improvements across parameter scales.

Additional error breakdown confirms that corrections are broadly distributed across all major error classes, indicating that HopChain does not merely patch narrow failure slices but strengthens robustness over diverse reasoning challenges. Figure 8

Figure 8: Error distribution among cases improved by multi-hop data, demonstrating substantial correction across perception, reasoning, logic, and object-level failures.

Theoretical and Practical Implications

HopChain's data-centric perspective on vision-language RLVR provides a scalable means to close the gap between multi-step, perceptually grounded reasoning required by real-world tasks and the surrogates typically emphasized in existing training corpora. The induction of explicit, instance-coupled chains typeset for RL training directly mitigates both shallow shortcut exploitation and compounding error propagation. This architecture- and model-agnostic signal is shown to be composable, yielding gains not only in image-based but also video-based tasks, highlighting its transfer properties.

Scalability is bounded by segmentation capability; thus, further extensions could augment the instance-anchoring strategy to handle non-segmentable or abstract images. This would permit generalized deployment in less structured or more abstract visual domains.

Conclusion

HopChain operationalizes multi-hop visual reasoning as a generative training paradigm that structurally targets compounding errors in long-CoT vision-LLMs. By synthesizing logically dependent, instance-anchored chains and integrating them through RLVR with verifiable reward supervision, it enables generalizable improvement—spanning domains, model scales, and error types—without reliance on benchmark overfitting. Its data-driven induction strategy offers a template for further advances in robust, scalable, and highly generalizable multimodal reasoning systems.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We found no open problems mentioned in this paper.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 8 tweets with 86 likes about this paper.