Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
113 tokens/sec
GPT-4o
12 tokens/sec
Gemini 2.5 Pro Pro
24 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
3 tokens/sec
DeepSeek R1 via Azure Pro
33 tokens/sec
2000 character limit reached

Transfer learning from language models to image caption generators: Better models may not transfer better (1901.01216v1)

Published 1 Jan 2019 in cs.CL, cs.LG, and cs.NE

Abstract: When designing a neural caption generator, a convolutional neural network can be used to extract image features. Is it possible to also use a neural LLM to extract sentence prefix features? We answer this question by trying different ways to transfer the recurrent neural network and embedding layer from a neural LLM to an image caption generator. We find that image caption generators with transferred parameters perform better than those trained from scratch, even when simply pre-training them on the text of the same captions dataset it will later be trained on. We also find that the best LLMs (in terms of perplexity) do not result in the best caption generators after transfer learning.

Citations (3)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.