MMG-Vid: Efficient Visual Token Pruning
- MMG-Vid is a training-free visual token pruning framework that reduces computational load in video LLMs by dynamically optimizing token selection at both segment and token levels.
- It segments videos based on frame similarity, allocates dynamic token budgets, and applies temporal-guided clustering to preserve the most informative and diverse tokens.
- MMG-Vid achieves over 99.5% performance retention while pruning 75% of tokens, significantly improving latency for real-world video-language applications.
MMG‑Vid is a training-free visual token pruning framework for Video LLMs (@@@@1@@@@) that addresses the computational challenges posed by excessive visual token counts during inference. MMG-Vid removes redundancy by dynamically maximizing marginal gains at both segment-level and token-level. The method prioritizes tokens based on their representativeness and diversity within semantic segments and over time, yielding substantial reductions in token count and inference latency with negligible performance degradation in video-language tasks (Ma et al., 28 Aug 2025).
1. Motivation and Framework Overview
VLLMs process video inputs by converting each frame into hundreds or thousands of visual tokens, feeding these into quadratic-complexity self-attention layers. This approach yields excellent video understanding, but the computational cost—especially at inference—limits deployment for real-world tasks where efficiency is paramount. Token pruning is essential, yet prior work either applies naive, static strategies or ignores video dynamics that introduce temporal redundancy.
MMG-Vid operates in three stages:
- Segment-Level Pruning: Videos are divided into meaningful segments based on frame similarity metrics.
- Dynamic Token Budget Allocation: Each segment receives a tailored number of tokens, determined by its marginal informational value.
- Token-Level Pruning with Temporal Guidance: Within each segment, a temporal-guided clustering algorithm preserves tokens that are locally and temporally salient.
This training-free, plug-and-play pipeline is agnostic to underlying VLLM architectures and requires no retraining.
2. Segment-Level Division and Dynamic Token Budgeting
Videos are first segmented based on frame-level feature similarity. Frame embeddings are compared with cosine similarity: Frames where this metric drops below a threshold become segment boundaries. To avoid overly granular segmentation, single-frame segments are merged with their most similar neighbor.
Each segment receives a dynamically apportioned token budget via a marginal gain formulation: where:
- is the segment embedding,
- is the average of the unselected segments,
- is the average of the already selected segments,
- controls representativeness vs. diversity.
Budgets are computed by Z-score normalization: This mechanism ensures complex, information-rich segments are prioritized with more tokens.
3. Token-Level Pruning via Temporal-Guided DPC
Within each segment, token selection considers both intra-frame saliency and inter-frame novelty.
For the first frame in a segment, the Density-Peak Clustering (DPC)-KNN approach is used:
- Compute local density for token :
- Determine minimum distance to higher-density tokens:
- Set token score , and select highest-scoring tokens.
For subsequent frames, the Temporal-Guided DPC (TG-DPC) algorithm uses reference tokens previously selected from earlier frames:
- Temporal relevance density: where are nearest tokens from previous frames.
- Temporal intra-frame separation is computed analogously to DPC, but within the current frame.
- New score:
Tokens are then selected to maximize temporal uniqueness and maintain intra-frame diversity, optimizing for utilization of a strict token budget.
4. Performance Evaluation and Metrics
The quality of a selected token subset is quantified as: where measures representativeness and quantifies redundancy.
On LLaVA-OneVision-7B, MMG-Vid retains >99.5% of the original VLLM performance even after pruning 75% of visual tokens. Prefilling latency is reduced by 3.9, and generation time by 3.1 relative to baseline. These improvements are consistent across video benchmarks such as MVBench, LongVideoBench, MLVU, and VideoMME.
5. Comparative Analysis and Ablation Studies
Ablation results establish the necessity of dynamic segment budgeting and temporal-guided token pruning. Replacing TG-DPC with standard frame-wise clustering or using uniform token allocation yields observable drops in performance. Compared to previous approaches (FastV, VisionZip, PruneVid, FrameFusion), MMG-Vid consistently demonstrates either equivalent or superior retention of accuracy under substantially lower token counts.
6. Implications for Video-LLM Deployment
MMG-Vid enables real-time or resource-constrained deployment of VLLMs by pushing the computational bottleneck down through strategic pruning, while maintaining semantic and temporal comprehensiveness in the video token representation. Its training-free and modular design makes it broadly applicable and easily integrated with existing VLLM architectures.
A plausible implication is that such efficient token selection will become standard for mobile, edge, or low-latency VLLM deployments, and may further encourage research into adaptive, context-aware pruning strategies. The framework’s design—jointly maximizing intra-segment and intra-frame marginal gains—represents a generalizable approach for redundancy reduction in sequential visual token processing.
7. Future Directions
The MMG-Vid framework suggests several avenues for further exploration:
- Enhancing temporal modeling by conditioning token selection on higher-order dynamics or longer-term dependencies.
- Extending pruning strategies to multi-modal fusion, for example in joint video-text architectures.
- Incorporating training-aware or differentiable pruning mechanisms, potentially improving end-to-end integration.
- Investigating how such strategies translate to other domains where sequential redundancy is an issue (e.g., long-form document LLMs, medical video analytics).
In conclusion, MMG-Vid maximizes efficiency and maintains fidelity for video LLMs in real-world applications by jointly optimizing segment-level and token-level marginal gains—a significant contribution to the deployment and scalability of video–language understanding systems (Ma et al., 28 Aug 2025).