Generalizing robustness across semantic and modality perturbations in multimodal LLMs
Develop mechanisms for Multimodal Large Language Models that generalize across both semantic perturbations (e.g., misleading or irrelevant content in text or images) and modality perturbations (e.g., input- or embedding-level noise), so that robustness is maintained without relying on task-specific or attack-specific defenses.
References
Developing mechanisms that generalize across both semantic and modality perturbations remains an open and challenging direction.
— Diagnosing and Mitigating Modality Interference in Multimodal Large Language Models
(2505.19616 - Cai et al., 26 May 2025) in Appendix, Section Limitations