Papers
Topics
Authors
Recent
Search
2000 character limit reached

Seg-ReSearch: Agentic Segmentation Framework

Updated 5 February 2026
  • Seg-ReSearch is a novel agentic segmentation paradigm that interleaves multimodal language model reasoning with external search to overcome static knowledge limitations.
  • It decomposes user instructions into multi-hop queries and integrates live external data to generate precise segmentation masks for up-to-date or domain-specific challenges.
  • The framework employs a hierarchical reward mechanism in reinforcement learning to balance process guidance with segmentation accuracy, validated on benchmarks like OK-VOS.

Seg-ReSearch defines a novel agentic paradigm for segmentation in computer vision by interleaving multimodal LLM (MLLM) reasoning with external search, directly addressing the knowledge bottleneck faced by previous approaches that relied solely on the frozen internal knowledge of their backbone LLMs. This framework enables segmentation systems to handle dynamic, open-world queries—including those involving up-to-date events or domain-specific knowledge—by decomposing user instructions, querying external sources, integrating retrieved evidence, and ultimately generating segmentation masks. The framework introduces a hierarchical reward mechanism to reconcile sparse outcome signals with flexible, process-level reinforcement, and is evaluated on the newly constructed OK-VOS benchmark as well as previous @@@@1@@@@ datasets (Liang et al., 4 Feb 2026).

1. Motivation and Challenge of Knowledge Bottlenecks

Recent MLLM-based segmentation architectures, such as LISA and VideoSeg-R1, have advanced multi-modal reasoning capabilities for instruction-based segmentation. However, these systems are fundamentally limited by their static knowledge bases, resulting in failures when queries reference facts or objects not included in their training data—for example, contemporary events or niche product launches. Real-world image and video-based queries frequently require access to dynamic and external knowledge, such as segmenting newly released objects or actors who achieved specific accolades in recent timeframes. Addressing such queries necessitates not only decomposing multi-hop instructions but also issuing live searches and integrating newfound evidence into the segmentation workflow, which prior SOTA methods cannot accomplish.

Seg-ReSearch implements an interleaved process between policy-driven reasoning and external search mechanisms, optimizing segmentation for open-world, dynamically evolving queries.

2.1 System Components

  • Segmentation Backbone (“Mask Generator”): A frozen, high-accuracy segmentation model (e.g., SAM2), accepting bounding-box and point prompts to generate final masks.
  • Policy MLLM (Reasoning Module): A trainable multimodal LLM that parses user queries and visual input, generates multi-turn chain-of-thought (MCoT) reasoning, and dynamically decides when and how to interact with external sources.
  • External Search Interface: Text and image search APIs returning the top-k relevant passages or images, providing the necessary external evidence for multi-hop or knowledge-intensive queries.
  • Reward Manager: Orchestrates a hierarchical reward structure for reinforcement learning.

2.2 Stepwise Interaction

At each interaction stage:

  1. Inputs consist of NN low-resolution frames {f1,...,fN}\{f_1, ..., f_N\} and a natural-language query.
  2. For timesteps t=1...Tt=1...T:
    • The policy MLLM analyzes the available context (past information blocks and frames).
    • When needed, it outputs \langlesearch\rangle blocks with search queries for text or images.
    • Results are returned and added to the context as \langleinformation\rangle blocks.
    • Visual-only reasoning is performed if external information is unnecessary.
  3. Upon sufficient confidence or after TT steps, a keyframe index kk is output via \langlekeyframe\rangle.
  4. The chosen high-resolution frame fkf_k is appended, and the model outputs an \langleanswer\rangle JSON with 2D bounding box and point prompts.
  5. These prompts drive the frozen backbone to produce the final segmentation mask.

This interleaved design enables the system to flexibly handle multi-hop, temporally sensitive, or domain-specific segmentation challenges.

3. Hierarchical Reward Mechanism and Formal Definitions

Seg-ReSearch is trained via reinforcement learning, employing a hierarchical reward RR that combines initial, process, and outcome-based signals:

R=α(RIGR+RTPR)+RORR = \alpha \cdot (R_{\mathrm{IGR}} + R_{\mathrm{TPR}}) + R_{\mathrm{OR}}

Where α[0,1]\alpha \in [0,1] determines the balance between process and outcome signals.

  • Initial Guidance Reward (RIGRR_{\mathrm{IGR}}): Binary signal for “warm-start” search:

RIGR=1{maxsSSim(a0,s)>0.5}R_{\mathrm{IGR}} = 1\{ \max_{s \in S} \mathrm{Sim}(a_0, s) > 0.5 \}

a0a_0: first policy search query; SS: expert queries; Sim\mathrm{Sim}: semantic similarity.

  • Tapering Process Reward (RTPRR_{\mathrm{TPR}}): Saturating bonus to encourage valid exploration but discourage unnecessary actions:

RTPR=1(1p)min(k,M)R_{\mathrm{TPR}} = 1 - (1 - p)^{\min(k, M)}

p[0,1]p\in[0,1]: base per-action reward, kk: valid actions, MM: bonus cap. Increased kk yields diminishing returns as RTPR1R_{\mathrm{TPR}} \rightarrow 1.

  • Outcome Reward (RORR_{\mathrm{OR}}): Composite of downstream segmentation accuracy and keyframe selection:

ROR=Riou+Rl1+Rpoint+RframeR_{\mathrm{OR}} = R_{\mathrm{iou}} + R_{\mathrm{l1}} + R_{\mathrm{point}} + R_{\mathrm{frame}}

Where Riou=1R_{\mathrm{iou}}=1 if IoU >0.5>0.5, Rl1=1R_{\mathrm{l1}}=1 if L1L1 distance <10<10 px, Rpoint=1R_{\mathrm{point}}=1 if the predicted point is present in the box and Euclidean distance <100<100 px, and RframeR_{\mathrm{frame}} scales with the largest

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Seg-ReSearch.