Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improving Captioning for Low-Resource Languages by Cycle Consistency (1908.07810v1)

Published 21 Aug 2019 in cs.CL and cs.MM

Abstract: Improving the captioning performance on low-resource languages by leveraging English caption datasets has received increasing research interest in recent years. Existing works mainly fall into two categories: translation-based and alignment-based approaches. In this paper, we propose to combine the merits of both approaches in one unified architecture. Specifically, we use a pre-trained English caption model to generate high-quality English captions, and then take both the image and generated English captions to generate low-resource language captions. We improve the captioning performance by adding the cycle consistency constraint on the cycle of image regions, English words, and low-resource language words. Moreover, our architecture has a flexible design which enables it to benefit from large monolingual English caption datasets. Experimental results demonstrate that our approach outperforms the state-of-the-art methods on common evaluation metrics. The attention visualization also shows that the proposed approach really improves the fine-grained alignment between words and image regions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yike Wu (13 papers)
  2. Shiwan Zhao (47 papers)
  3. Jia Chen (85 papers)
  4. Ying Zhang (389 papers)
  5. Xiaojie Yuan (26 papers)
  6. Zhong Su (8 papers)
Citations (7)

Summary

We haven't generated a summary for this paper yet.