Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

How Powerful Potential of Attention on Image Restoration? (2403.10336v1)

Published 15 Mar 2024 in cs.CV

Abstract: Transformers have demonstrated their effectiveness in image restoration tasks. Existing Transformer architectures typically comprise two essential components: multi-head self-attention and feed-forward network (FFN). The former captures long-range pixel dependencies, while the latter enables the model to learn complex patterns and relationships in the data. Previous studies have demonstrated that FFNs are key-value memories \cite{geva2020transformer}, which are vital in modern Transformer architectures. In this paper, we conduct an empirical study to explore the potential of attention mechanisms without using FFN and provide novel structures to demonstrate that removing FFN is flexible for image restoration. Specifically, we propose Continuous Scaling Attention (\textbf{CSAttn}), a method that computes attention continuously in three stages without using FFN. To achieve competitive performance, we propose a series of key components within the attention. Our designs provide a closer look at the attention mechanism and reveal that some simple operations can significantly affect the model performance. We apply our \textbf{CSAttn} to several image restoration tasks and show that our model can outperform CNN-based and Transformer-based image restoration approaches.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Cong Wang (310 papers)
  2. Jinshan Pan (80 papers)
  3. Yeying Jin (29 papers)
  4. Liyan Wang (7 papers)
  5. Wei Wang (1793 papers)
  6. Gang Fu (19 papers)
  7. Wenqi Ren (67 papers)
  8. Xiaochun Cao (177 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.