Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Deformably-Scaled Transposed Convolution (2210.09446v1)

Published 17 Oct 2022 in cs.CV, cs.AI, cs.LG, and q-bio.NC

Abstract: Transposed convolution is crucial for generating high-resolution outputs, yet has received little attention compared to convolution layers. In this work we revisit transposed convolution and introduce a novel layer that allows us to place information in the image selectively and choose the `stroke breadth' at which the image is synthesized, whilst incurring a small additional parameter cost. For this we introduce three ideas: firstly, we regress offsets to the positions where the transpose convolution results are placed; secondly we broadcast the offset weight locations over a learnable neighborhood; and thirdly we use a compact parametrization to share weights and restrict offsets. We show that simply substituting upsampling operators with our novel layer produces substantial improvements across tasks as diverse as instance segmentation, object detection, semantic segmentation, generative image modeling, and 3D magnetic resonance image enhancement, while outperforming all existing variants of transposed convolutions. Our novel layer can be used as a drop-in replacement for 2D and 3D upsampling operators and the code will be publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Stefano B. Blumberg (18 papers)
  2. Mou-Cheng Xu (8 papers)
  3. Matteo Figini (10 papers)
  4. Iasonas Kokkinos (38 papers)
  5. Daniel C. Alexander (81 papers)
  6. Daniele RavĂ­ (1 paper)
Citations (1)