ProRes: Exploring Degradation-aware Visual Prompt for Universal Image Restoration (2306.13653v1)
Abstract: Image restoration aims to reconstruct degraded images, e.g., denoising or deblurring. Existing works focus on designing task-specific methods and there are inadequate attempts at universal methods. However, simply unifying multiple tasks into one universal architecture suffers from uncontrollable and undesired predictions. To address those issues, we explore prompt learning in universal architectures for image restoration tasks. In this paper, we present Degradation-aware Visual Prompts, which encode various types of image degradation, e.g., noise and blur, into unified visual prompts. These degradation-aware prompts provide control over image processing and allow weighted combinations for customized image restoration. We then leverage degradation-aware visual prompts to establish a controllable and universal model for image restoration, called ProRes, which is applicable to an extensive range of image restoration tasks. ProRes leverages the vanilla Vision Transformer (ViT) without any task-specific designs. Furthermore, the pre-trained ProRes can easily adapt to new tasks through efficient prompt tuning with only a few images. Without bells and whistles, ProRes achieves competitive performance compared to task-specific methods and experiments can demonstrate its ability for controllable restoration and adaptation for new tasks. The code and models will be released in \url{https://github.com/leonmakise/ProRes}.
- Simple baselines for image restoration. In ECCV, pages 17--33, 2022.
- Learning enriched features for fast image restoration and enhancement. IEEE Trans. Pattern Anal. Mach. Intell., 45(2):1934--1948, 2023.
- Mprnet: Multi-path residual network for lightweight image super resolution. In WACV, pages 2703--2712, 2021.
- Uformer: A general u-shaped transformer for image restoration. In CVPR, pages 17662--17672, 2022.
- All-in-one image restoration for unknown corruption. In CVPR, pages 17431--17441, 2022.
- Learning multiple adverse weather removal via two-stage knowledge learning and multi-contrastive regularization: Toward a unified model. In CVPR, pages 17632--17641, 2022.
- Path-restore: Learning network path selection for image restoration. IEEE Trans. Pattern Anal. Mach. Intell., 44(10):7078--7092, 2022.
- Uni-perceiver-moe: Learning sparse generalist models with conditional moes. In NeurIPS, 2022.
- Uni-perceiver: Pre-training unified architecture for generic perception for zero-shot and few-shot tasks. In CVPR, 2022.
- Masked-attention mask transformer for universal image segmentation. In CVPR, 2022.
- Oneformer: One transformer to rule universal image segmentation. In CoRR, volume abs/2211.06220, 2022.
- An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2021.
- A high-quality denoising dataset for smartphone cameras. In CVPR, pages 1692--1700, 2018.
- Deep retinex decomposition for low-light enhancement. In BMVC, pages 1--12, 2018.
- BERT: pre-training of deep bidirectional transformers for language understanding. In NAACL, pages 4171--4186, 2019.
- Perceiver: General perception with iterative attention. In ICML, pages 4651--4664, 2021.
- Perceiver IO: A general architecture for structured inputs & outputs. In ICLR, 2022.
- Uni-perceiver: Pre-training unified architecture for generic perception for zero-shot and few-shot tasks. In CVPR, pages 16783--16794, 2022.
- UPGPT: universal diffusion model for person image generation, editing and pose transfer. CoRR, abs/2304.08870, 2023.
- Universal cross-domain 3d model retrieval. IEEE Trans. Multim., 23:2721--2731, 2021.
- Could giant pretrained image models extract universal representations? CoRR, abs/2211.02043, 2022.
- Towards A universal model for cross-dataset crowd counting. In ICCV, pages 3185--3194, 2021.
- Visual prompting via image inpainting. In NeurIPS, 2022.
- Images speak in images: A generalist painter for in-context visual learning. CoRR, abs/2212.02499, 2022.
- Learning transferable visual models from natural language supervision. In ICML, volume 139, pages 8748--8763, 2021.
- Open compound domain adaptation. In CVPR, pages 12807--12816, 2020.
- Align before fuse: Vision and language representation learning with momentum distillation. In CVPR, pages 12727--12737, 2021.
- Masked autoencoders are scalable vision learners. In CVPR, 2022.
- Deep retinex decomposition for low-light enhancement. In BMVC, page 155, 2018.
- Multi-stage progressive image restoration. In CVPR, pages 14821--14831, 2021.
- Deep multi-scale convolutional neural network for dynamic scene deblurring. In CVPR, pages 3883--3891, 2017.
- Hide: A hierarchical image dataset for deblurring. In CVPRW, 2019.
- Real-world blind image deblurring using an adaptive activation function. In ECCV, pages 3--19, 2020.
- Learning enriched features for real image restoration and enhancement. In ECCV, pages 3--19, 2020.
- Deep joint rain detection and removal from a single image. In CVPR, pages 1357--1366, 2017.
- Density-aware single image de-raining using a multi-stream dense network. In CVPR, pages 695--703, 2018.
- Multi-stage progressive image restoration. In CVPR, pages 2616--2625, 2021.
- Decoupled weight decay regularization. In ICLR, 2019.
- Deep networks with stochastic depth. In ECCV, pages 646--661, 2016.
- Restormer: Efficient transformer for high-resolution image restoration. In CVPR, pages 5718--5729, 2022.
- MAXIM: multi-axis MLP for image processing. In CVPR, pages 5759--5770, 2022.
- Learning photographic global tonal adjustment with a database of input / output image pairs. In CVPR, pages 97--104, 2011.
- NH-HAZE: an image dehazing benchmark with non-homogeneous hazy and haze-free images. In CVPR, pages 1798--1805, 2020.
- Two deterministic half-quadratic regularization algorithms for computed imaging. In ICIP, pages 168--172, 1994.
- Perceptual losses for real-time style transfer and super-resolution. In ECCV, volume 9906, pages 694--711, 2016.
- Jiaqi Ma (83 papers)
- Tianheng Cheng (31 papers)
- Guoli Wang (40 papers)
- Qian Zhang (308 papers)
- Xinggang Wang (163 papers)
- Lefei Zhang (64 papers)