Accurate Word Representations with Universal Visual Guidance (2012.15086v1)
Abstract: Word representation is a fundamental component in neural language understanding models. Recently, pre-trained LLMs (PrLMs) offer a new performant method of contextualized word representations by leveraging the sequence-level context for modeling. Although the PrLMs generally give more accurate contextualized word representations than non-contextualized models do, they are still subject to a sequence of text contexts without diverse hints for word representation from multimodality. This paper thus proposes a visual representation method to explicitly enhance conventional word embedding with multiple-aspect senses from visual guidance. In detail, we build a small-scale word-image dictionary from a multimodal seed dataset where each word corresponds to diverse related images. The texts and paired images are encoded in parallel, followed by an attention layer to integrate the multimodal representations. We show that the method substantially improves the accuracy of disambiguation. Experiments on 12 natural language understanding and machine translation tasks further verify the effectiveness and the generalization capability of the proposed approach.
- Zhuosheng Zhang (125 papers)
- Haojie Yu (4 papers)
- Hai Zhao (227 papers)
- Rui Wang (996 papers)
- Masao Utiyama (39 papers)