Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

VLDeformer: Vision-Language Decomposed Transformer for Fast Cross-Modal Retrieval (2110.11338v3)

Published 20 Oct 2021 in cs.CV, cs.CL, and cs.IR

Abstract: Cross-model retrieval has emerged as one of the most important upgrades for text-only search engines (SE). Recently, with powerful representation for pairwise text-image inputs via early interaction, the accuracy of vision-language (VL) transformers has outperformed existing methods for text-image retrieval. However, when the same paradigm is used for inference, the efficiency of the VL transformers is still too low to be applied in a real cross-modal SE. Inspired by the mechanism of human learning and using cross-modal knowledge, this paper presents a novel Vision-Language Decomposed Transformer (VLDeformer), which greatly increases the efficiency of VL transformers while maintaining their outstanding accuracy. By the proposed method, the cross-model retrieval is separated into two stages: the VL transformer learning stage, and the VL decomposition stage. The latter stage plays the role of single modal indexing, which is to some extent like the term indexing of a text SE. The model learns cross-modal knowledge from early-interaction pre-training and is then decomposed into an individual encoder. The decomposition requires only small target datasets for supervision and achieves both $1000+$ times acceleration and less than $0.6$\% average recall drop. VLDeformer also outperforms state-of-the-art visual-semantic embedding methods on COCO and Flickr30k.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Lisai Zhang (8 papers)
  2. Hongfa Wu (1 paper)
  3. Qingcai Chen (36 papers)
  4. Yimeng Deng (1 paper)
  5. Zhonghua Li (46 papers)
  6. Dejiang Kong (2 papers)
  7. Zhao Cao (36 papers)
  8. Joanna Siebert (5 papers)
  9. Yunpeng Han (4 papers)
Citations (17)