Breadth–Depth Compute Allocation for LVLM Test-Time Reasoning
Determine the optimal allocation of test-time compute between breadth (sampling more reasoning paths via multi-pass decoding) and depth (using stronger chain-of-thought or "thinking" modes) for large vision–language models on perception tasks.
References
We do not yet know how to best allocate test-time compute between sampling more reasoning paths (breadth) and using stronger reasoning modes (depth) in perception tasks.
— When to Think and When to Look: Uncertainty-Guided Lookback
(2511.15613 - Bi et al., 19 Nov 2025) in Section 1 (Introduction), under the question "How should we trade off breadth vs. depth of thinking?"