Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SPA-GAN: Spatial Attention GAN for Image-to-Image Translation (1908.06616v3)

Published 19 Aug 2019 in cs.CV

Abstract: Image-to-image translation is to learn a mapping between images from a source domain and images from a target domain. In this paper, we introduce the attention mechanism directly to the generative adversarial network (GAN) architecture and propose a novel spatial attention GAN model (SPA-GAN) for image-to-image translation tasks. SPA-GAN computes the attention in its discriminator and use it to help the generator focus more on the most discriminative regions between the source and target domains, leading to more realistic output images. We also find it helpful to introduce an additional feature map loss in SPA-GAN training to preserve domain specific features during translation. Compared with existing attention-guided GAN models, SPA-GAN is a lightweight model that does not need additional attention networks or supervision. Qualitative and quantitative comparison against state-of-the-art methods on benchmark datasets demonstrates the superior performance of SPA-GAN.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Hajar Emami (5 papers)
  2. Majid Moradi Aliabadi (1 paper)
  3. Ming Dong (38 papers)
  4. Ratna Babu Chinnam (9 papers)
Citations (150)

Summary

We haven't generated a summary for this paper yet.