Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 93 tok/s
Gemini 2.5 Pro 55 tok/s Pro
GPT-5 Medium 25 tok/s
GPT-5 High 24 tok/s Pro
GPT-4o 91 tok/s
GPT OSS 120B 462 tok/s Pro
Kimi K2 209 tok/s Pro
2000 character limit reached

InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency (2508.18265v2)

Published 25 Aug 2025 in cs.CV

Abstract: We introduce InternVL 3.5, a new family of open-source multimodal models that significantly advances versatility, reasoning capability, and inference efficiency along the InternVL series. A key innovation is the Cascade Reinforcement Learning (Cascade RL) framework, which enhances reasoning through a two-stage process: offline RL for stable convergence and online RL for refined alignment. This coarse-to-fine training strategy leads to substantial improvements on downstream reasoning tasks, e.g., MMMU and MathVista. To optimize efficiency, we propose a Visual Resolution Router (ViR) that dynamically adjusts the resolution of visual tokens without compromising performance. Coupled with ViR, our Decoupled Vision-Language Deployment (DvD) strategy separates the vision encoder and LLM across different GPUs, effectively balancing computational load. These contributions collectively enable InternVL3.5 to achieve up to a +16.0\% gain in overall reasoning performance and a 4.05$\times$ inference speedup compared to its predecessor, i.e., InternVL3. In addition, InternVL3.5 supports novel capabilities such as GUI interaction and embodied agency. Notably, our largest model, i.e., InternVL3.5-241B-A28B, attains state-of-the-art results among open-source MLLMs across general multimodal, reasoning, text, and agentic tasks -- narrowing the performance gap with leading commercial models like GPT-5. All models and code are publicly released.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces a modular ViT–MLP–LLM architecture featuring a Visual Resolution Router that adaptively compresses visual tokens with minimal performance loss.
  • The Cascade RL training paradigm, which combines offline MPO and online GSPO, yields up to a 16% improvement in reasoning benchmarks and enhances training efficiency.
  • The Decoupled Vision-Language Deployment framework, paired with ViR, enables up to 4.05× faster inference, supporting both dense and MoE configurations in resource-constrained settings.

InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency

Model Architecture and Innovations

InternVL3.5 extends the InternVL series with a modular "ViT–MLP–LLM" architecture, leveraging InternViT vision encoders and Qwen3/GPT-OSS LLMs. The model supports both dense and MoE configurations, scaling from 1B to 241B parameters. A key architectural innovation is the Visual Resolution Router (ViR), which adaptively compresses visual tokens based on semantic richness, enabling dynamic trade-offs between computational cost and representational fidelity. This is integrated via a pixel shuffle module, reducing the number of visual tokens by up to 50% with negligible performance loss. Figure 1

Figure 1: InternVL3.5 architecture with ViR for adaptive visual token compression, supporting efficient multimodal processing.

InternVL3.5-Flash variants incorporate ViR, making them suitable for resource-constrained deployments. The Decoupled Vision-Language Deployment (DvD) framework further optimizes inference by parallelizing vision and language processing across separate servers, maximizing hardware utilization and throughput. Figure 2

Figure 2: DvD enables parallel, asynchronous execution of vision and language modules, reducing inference latency and resource contention.

Training Paradigm: Cascade Reinforcement Learning

InternVL3.5 introduces Cascade RL, a two-stage RL framework combining offline Mixed Preference Optimization (MPO) and online Generalized Self-Play Optimization (GSPO). The offline stage (MPO) efficiently warms up the model using preference, quality, and generation losses, while the online stage (GSPO) refines output distributions via normalized rewards and importance sampling. This coarse-to-fine approach yields stable convergence, high-quality rollouts, and a higher performance ceiling compared to single-stage RL paradigms. Figure 3

Figure 3: Training pipeline: native pre-training, supervised fine-tuning, Cascade RL, and ViR integration via consistency and router training.

Cascade RL demonstrates strong scalability, with consistent gains from 1B to 241B models. Ablation studies show up to +16.0% improvement in reasoning benchmarks and superior training efficiency—Cascade RL achieves higher performance than GSPO with half the GPU hours. Figure 4

Figure 4: Cascade RL ablation: progressive gains in multimodal reasoning across model scales and training stages.

Efficiency: Visual Resolution Router and Decoupled Deployment

ViR is trained via Visual Consistency Learning (ViCO), minimizing KL divergence between outputs conditioned on different compression rates. Router training uses cross-entropy loss, with dynamic thresholds based on historical loss ratios. ViR enables InternVL3.5-Flash to maintain nearly 100% of original performance while halving visual token count.

DvD decouples vision and language inference, batching vision features and asynchronously transmitting them to the language server. This design yields up to 4.05× speedup in high-resolution scenarios and scales efficiently with input size and model capacity.

Empirical Results and Benchmarking

InternVL3.5 achieves state-of-the-art results among open-source MLLMs across 35 multimodal benchmarks, including general, reasoning, text, agentic, OCR, chart, document, multi-image, real-world, hallucination, grounding, multilingual, video, GUI, embodied, and SVG tasks. Figure 5

Figure 5: InternVL3.5 outperforms leading open-source and commercial MLLMs across multimodal, reasoning, text, and agentic benchmarks.

  • General Multimodal: InternVL3.5-241B-A28B matches GPT-5 on overall score (74.1 vs. 74.0), with strong results on MMVet, MMBench, and LongVideoBench.
  • Reasoning: Cascade RL yields scores of 77.7 (MMMU) and 82.7 (MathVista), surpassing open-source baselines and approaching commercial leaders.
  • Text: Native pre-training and Cascade RL preserve and enhance language capabilities, outperforming base LLMs and mitigating catastrophic forgetting.
  • Agentic/GUI: InternVL3.5 excels in GUI grounding and agentic tasks, outperforming specialized and generalist models on ScreenSpot, OSWorld-G, and WindowsAgentArena.
  • Embodied/SVG: InternVL3.5 sets new open-source records on VSI-Bench, SGP-Bench, and SArena-Icon, with competitive SVG generation metrics (FID, SSIM, LPIPS).
  • Efficiency: ViR and DvD together yield up to 4.05× inference speedup with minimal accuracy loss, validated across dense and MoE models.

Implementation and Deployment Considerations

  • Training: XTuner framework with FSDP, data packing, FP8 training, FlashAttention-3, and TMA-Adaptive FP8 Grouped GEMM for scalable LLM/MoE optimization.
  • Inference: DvD supports asynchronous, hardware-optimized deployment; ViR enables adaptive token compression for real-time applications.
  • Resource Requirements: Cascade RL is GPU-efficient, with offline RL amortizing rollout costs and online RL leveraging high-quality rollouts for robust convergence.
  • Scaling: InternVL3.5 demonstrates linear scalability in reasoning and efficiency across model sizes, with consistent performance improvements in both dense and MoE configurations.

Implications and Future Directions

InternVL3.5 narrows the gap between open-source and commercial MLLMs, particularly in reasoning, agentic, and efficiency domains. The Cascade RL framework provides a scalable, stable RL paradigm for multimodal models, while ViR and DvD offer practical solutions for real-world deployment bottlenecks. The model's versatility across text, vision, video, GUI, embodied, and SVG tasks positions it as a foundational platform for multimodal AI research and applications.

Future work may focus on further reducing hallucination rates, enhancing robustness in long-context and multi-image/video scenarios, and extending agentic capabilities for embodied and interactive environments. The open-source release of models and code will facilitate reproducibility, benchmarking, and community-driven innovation in multimodal AI.

Conclusion

InternVL3.5 advances open-source multimodal modeling through architectural, training, and deployment innovations. Cascade RL, ViR, and DvD collectively yield substantial gains in reasoning, versatility, and efficiency, with empirical results demonstrating state-of-the-art performance across diverse benchmarks. The model's scalability, efficiency, and broad applicability make it a significant contribution to the multimodal AI landscape, providing a robust foundation for future research and deployment.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Reddit Logo Streamline Icon: https://streamlinehq.com

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube