Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

TextTopicNet - Self-Supervised Learning of Visual Features Through Embedding Images on Semantic Text Spaces (1807.02110v1)

Published 4 Jul 2018 in cs.CV

Abstract: The immense success of deep learning based methods in computer vision heavily relies on large scale training datasets. These richly annotated datasets help the network learn discriminative visual features. Collecting and annotating such datasets requires a tremendous amount of human effort and annotations are limited to popular set of classes. As an alternative, learning visual features by designing auxiliary tasks which make use of freely available self-supervision has become increasingly popular in the computer vision community. In this paper, we put forward an idea to take advantage of multi-modal context to provide self-supervision for the training of computer vision algorithms. We show that adequate visual features can be learned efficiently by training a CNN to predict the semantic textual context in which a particular image is more probable to appear as an illustration. More specifically we use popular text embedding techniques to provide the self-supervision for the training of deep CNN. Our experiments demonstrate state-of-the-art performance in image classification, object detection, and multi-modal retrieval compared to recent self-supervised or naturally-supervised approaches.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yash Patel (41 papers)
  2. Lluis Gomez (42 papers)
  3. Raul Gomez (16 papers)
  4. Marçal Rusiñol (20 papers)
  5. Dimosthenis Karatzas (80 papers)
  6. C. V. Jawahar (110 papers)
Citations (7)