Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Lessons learned in multilingual grounded language learning (1809.07615v1)

Published 20 Sep 2018 in cs.CL

Abstract: Recent work has shown how to learn better visual-semantic embeddings by leveraging image descriptions in more than one language. Here, we investigate in detail which conditions affect the performance of this type of grounded language learning model. We show that multilingual training improves over bilingual training, and that low-resource languages benefit from training with higher-resource languages. We demonstrate that a multilingual model can be trained equally well on either translations or comparable sentence pairs, and that annotating the same set of images in multiple language enables further improvements via an additional caption-caption ranking objective.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Desmond Elliott (53 papers)
  2. Marc-Alexandre Côté (42 papers)
  3. Grzegorz Chrupała (33 papers)
  4. Afra Alishahi (26 papers)
  5. Ákos Kádár (15 papers)
Citations (22)