Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Spatial Transformer Networks for Curriculum Learning (2108.09696v1)

Published 22 Aug 2021 in cs.CV

Abstract: Curriculum learning is a bio-inspired training technique that is widely adopted to machine learning for improved optimization and better training of neural networks regarding the convergence rate or obtained accuracy. The main concept in curriculum learning is to start the training with simpler tasks and gradually increase the level of difficulty. Therefore, a natural question is how to determine or generate these simpler tasks. In this work, we take inspiration from Spatial Transformer Networks (STNs) in order to form an easy-to-hard curriculum. As STNs have been proven to be capable of removing the clutter from the input images and obtaining higher accuracy in image classification tasks, we hypothesize that images processed by STNs can be seen as easier tasks and utilized in the interest of curriculum learning. To this end, we study multiple strategies developed for shaping the training curriculum, using the data generated by STNs. We perform various experiments on cluttered MNIST and Fashion-MNIST datasets, where on the former, we obtain an improvement of $3.8$pp in classification accuracy compared to the baseline.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Fatemeh Azimi (5 papers)
  2. Jean-Francois Jacques Nicolas Nies (1 paper)
  3. Sebastian Palacio (17 papers)
  4. Federico Raue (33 papers)
  5. Jörn Hees (28 papers)
  6. Andreas Dengel (188 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.