Papers
Topics
Authors
Recent
Search
2000 character limit reached

mPLUG-2: A Modularized Multi-modal Foundation Model Across Text, Image and Video

Published 1 Feb 2023 in cs.CV, cs.CL, and cs.MM | (2302.00402v1)

Abstract: Recent years have witnessed a big convergence of language, vision, and multi-modal pretraining. In this work, we present mPLUG-2, a new unified paradigm with modularized design for multi-modal pretraining, which can benefit from modality collaboration while addressing the problem of modality entanglement. In contrast to predominant paradigms of solely relying on sequence-to-sequence generation or encoder-based instance discrimination, mPLUG-2 introduces a multi-module composition network by sharing common universal modules for modality collaboration and disentangling different modality modules to deal with modality entanglement. It is flexible to select different modules for different understanding and generation tasks across all modalities including text, image, and video. Empirical study shows that mPLUG-2 achieves state-of-the-art or competitive results on a broad range of over 30 downstream tasks, spanning multi-modal tasks of image-text and video-text understanding and generation, and uni-modal tasks of text-only, image-only, and video-only understanding. Notably, mPLUG-2 shows new state-of-the-art results of 48.0 top-1 accuracy and 80.3 CIDEr on the challenging MSRVTT video QA and video caption tasks with a far smaller model size and data scale. It also demonstrates strong zero-shot transferability on vision-language and video-language tasks. Code and models will be released in https://github.com/alibaba/AliceMind.

Citations (127)

Summary

  • The paper presents a modular design that integrates universal and modality-specific modules to balance collaboration and entanglement across modalities.
  • It employs a unified encoder-decoder architecture to support over 30 tasks, including video QA and captioning, using specialized vision and text components.
  • The study demonstrates significant improvements with a top-1 accuracy of 48.0 on MSRVTT video QA and an 80.3 CIDEr score on video caption tasks, underscoring robust zero-shot transferability.

mPLUG-2: A Modularized Multi-modal Foundation Model Across Text, Image, and Video

The paper introduces mPLUG-2, a multi-modal foundation model designed to integrate text, image, and video processing within a unified framework. Unlike traditional models focusing on monolithic architectures, mPLUG-2 emphasizes modularization, thus enhancing flexibility and adaptability of tasks across multiple modalities. The study outlines its primary contributions, detailing a modular network design aiming to balance modality collaboration and entanglement.

Key Contributions

  1. Modular Design: mPLUG-2 leverages a novel architecture, incorporating universal and modality-specific modules. This design addresses modality entanglement while encouraging collaboration, allowing the model to generalize across a wide array of tasks.
  2. Unified Approach: The model supports uni-modal and cross-modal applications through a shared encoder-decoder architecture, incorporating specialized components such as dual-vision encoders for images and video, and text encoders. This strengthens mPLUG-2's ability to adapt to diverse data types and objectives.
  3. Empirical Evaluation: Tested on over 30 tasks, mPLUG-2 achieved state-of-the-art results across multiple benchmarks, including top-1 accuracy and CIDEr scores in challenging video QA and caption tasks. This highlights its competitive edge despite a smaller scale compared to counterparts.
  4. Zero-shot Transferability: The study demonstrates robust zero-shot capabilities across vision-language and video-language tasks, which signifies potential for application in scenarios lacking domain-specific training data.

Numerical Results and Implications

mPLUG-2's empirical success is notable; it achieves a top-1 accuracy of 48.0 on MSRVTT video QA and 80.3 CIDEr on video caption tasks, showcasing significant improvements compared to previous models. Moreover, it maintains competitive performance on visual grounding, VQA, and image-text retrieval tasks. These results emphasize mPLUG-2's broad applicability and efficiency in resource usage.

Future Directions

The paper proposes further exploration into scaling mPLUG-2, expanding its modular components to handle additional modalities and tasks. Additionally, there is potential to integrate advanced attention mechanisms or incorporate external knowledge bases to enhance the contextual understanding of complex multi-modal tasks.

Conclusion

mPLUG-2 represents a significant step forward in multi-modal AI research, offering a flexible and efficient architecture that balances collaboration and entanglement across diverse modalities. By achieving state-of-the-art results in numerous tasks, it sets a precedent for future research and development in the domain of integrated multi-modal machine learning platforms. The availability of its codebase also encourages further experimentation and validation by the research community.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.