Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
GPT-4o
Gemini 2.5 Pro Pro
o3 Pro
GPT-4.1 Pro
DeepSeek R1 via Azure Pro
2000 character limit reached

Generalizing to Out-of-Sample Degradations via Model Reprogramming (2403.05886v1)

Published 9 Mar 2024 in cs.CV

Abstract: Existing image restoration models are typically designed for specific tasks and struggle to generalize to out-of-sample degradations not encountered during training. While zero-shot methods can address this limitation by fine-tuning model parameters on testing samples, their effectiveness relies on predefined natural priors and physical models of specific degradations. Nevertheless, determining out-of-sample degradations faced in real-world scenarios is always impractical. As a result, it is more desirable to train restoration models with inherent generalization ability. To this end, this work introduces the Out-of-Sample Restoration (OSR) task, which aims to develop restoration models capable of handling out-of-sample degradations. An intuitive solution involves pre-translating out-of-sample degradations to known degradations of restoration models. However, directly translating them in the image space could lead to complex image translation issues. To address this issue, we propose a model reprogramming framework, which translates out-of-sample degradations by quantum mechanic and wave functions. Specifically, input images are decoupled as wave functions of amplitude and phase terms. The translation of out-of-sample degradation is performed by adapting the phase term. Meanwhile, the image content is maintained and enhanced in the amplitude term. By taking these two terms as inputs, restoration models are able to handle out-of-sample degradations without fine-tuning. Through extensive experiments across multiple evaluation cases, we demonstrate the effectiveness and flexibility of our proposed framework. Our codes are available at \href{https://github.com/ddghjikle/Out-of-sample-restoration}{Github}.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (57)
  1. J. Liang, J. Cao, G. Sun, K. Zhang, L. Van Gool, and R. Timofte, “Swinir: Image restoration using swin transformer,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1833–1844, 2021.
  2. S. W. Zamir, A. Arora, S. Khan, M. Hayat, F. S. Khan, and M.-H. Yang, “Restormer: Efficient transformer for high-resolution image restoration,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5728–5739, 2022.
  3. X. Zhang, R. Jiang, T. Wang, and W. Luo, “Single image dehazing via dual-path recurrent network,” IEEE Transactions on Image Processing, vol. 30, pp. 5211–5222, 2021.
  4. L. Liu, L. Xie, X. Zhang, S. Yuan, X. Chen, W. Zhou, H. Li, and Q. Tian, “Tape: Task-agnostic prior embedding for image restoration,” arXiv preprint arXiv:2203.06074, 2022.
  5. S. Zhao, L. Zhang, Y. Shen, and Y. Zhou, “Refinednet: A weakly supervised refinement framework for single image dehazing,” IEEE Transactions on Image Processing, vol. 30, pp. 3391–3404, 2021.
  6. B. Li, Y. Gou, J. Z. Liu, H. Zhu, J. T. Zhou, and X. Peng, “Zero-shot image dehazing,” IEEE Transactions on Image Processing, vol. 29, pp. 8457–8466, 2020.
  7. B. Li, Y. Gou, S. Gu, J. Z. Liu, J. T. Zhou, and X. Peng, “You only look yourself: Unsupervised and untrained single image dehazing neural network,” International Journal of Computer Vision, vol. 129, no. 5, pp. 1754–1767, 2021.
  8. B. Li, X. Peng, Z. Wang, J. Xu, and D. Feng, “Aod-net: All-in-one dehazing network,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4770–4778, 2017.
  9. M. Long, Y. Cao, J. Wang, and M. Jordan, “Learning transferable features with deep adaptation networks,” in International Conference on Machine Learning, pp. 97–105.   PMLR, 2015.
  10. K. Saito, K. Watanabe, Y. Ushiku, and T. Harada, “Maximum classifier discrepancy for unsupervised domain adaptation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3723–3732, 2018.
  11. A. Lengyel, S. Garg, M. Milford, and J. C. van Gemert, “Zero-shot day-night domain adaptation with a physics prior,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4399–4409, 2021.
  12. S. Li, M. Xie, F. Lv, C. H. Liu, J. Liang, C. Qin, and W. Li, “Semantic concentration for domain adaptation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9102–9111, 2021.
  13. Z. Han, H. Sun, and Y. Yin, “Learning transferable parameters for unsupervised domain adaptation,” IEEE Transactions on Image Processing, 2022.
  14. Y. Tang, K. Han, J. Guo, C. Xu, Y. Li, C. Xu, and Y. Wang, “An image patch is a wave: Phase-aware vision mlp,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10 935–10 944, 2022.
  15. J.-J. Huang and P. L. Dragotti, “Winnet: Wavelet-inspired invertible network for image denoising,” IEEE Transactions on Image Processing, vol. 31, pp. 4377–4392, 2022.
  16. Y. Qu, Y. Chen, J. Huang, and Y. Xie, “Enhanced pix2pix dehazing network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8160–8168, 2019.
  17. C. Chen and H. Li, “Robust representation learning with feedback for single image deraining,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7742–7751, 2021.
  18. B. Li, W. Ren, D. Fu, D. Tao, D. Feng, W. Zeng, and Z. Wang, “Benchmarking single-image dehazing and beyond,” IEEE Transactions on Image Processing, vol. 28, no. 1, pp. 492–505, 2018.
  19. X. Liu, Y. Ma, Z. Shi, and J. Chen, “Griddehazenet: Attention-based multi-scale network for image dehazing,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 7314–7323, 2019.
  20. Y. Wang, J. Yu, and J. Zhang, “Zero-shot image restoration using denoising diffusion null-space model,” arXiv preprint arXiv:2212.00490, 2022.
  21. L. Van der Maaten and G. Hinton, “Visualizing data using t-sne,” Journal of Machine Learning Research, vol. 9, no. 11, 2008.
  22. J. Johnson, A. Alahi, and L. Fei-Fei, “Perceptual losses for real-time style transfer and super-resolution,” in European Conference on Computer Vision, pp. 694–711.   Springer, 2016.
  23. X. Chen, S. Wang, M. Long, and J. Wang, “Transferability vs. discriminability: Batch spectral penalization for adversarial domain adaptation,” in International Conference on Machine Learning, pp. 1081–1090.   PMLR, 2019.
  24. P.-Y. Chen, “Model reprogramming: Resource-efficient cross-domain machine learning,” arXiv preprint arXiv:2202.10629, 2022.
  25. G. F. Elsayed, I. Goodfellow, and J. Sohl-Dickstein, “Adversarial reprogramming of neural networks,” arXiv preprint arXiv:1806.11146, 2018.
  26. Y.-Y. Tsai, P.-Y. Chen, and T.-Y. Ho, “Transfer learning without knowing: Reprogramming black-box machine learning models with scarce data and limited resources,” in International Conference on Machine Learning, pp. 9614–9624.   PMLR, 2020.
  27. I. Melnyk, V. Chenthamarakshan, P.-Y. Chen, P. Das, A. Dhurandhar, I. Padhi, and D. Das, “Reprogramming large pretrained language models for antibody sequence infilling,” arXiv preprint arXiv:2210.07144, 2022.
  28. P. Neekhara, S. Hussain, J. Du, S. Dubnov, F. Koushanfar, and J. McAuley, “Cross-modal adversarial reprogramming,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 2427–2435, 2022.
  29. C.-H. H. Yang, Y.-Y. Tsai, and P.-Y. Chen, “Voice2series: Reprogramming acoustic models for time series classification,” in International Conference on Machine Learning, pp. 11 808–11 819.   PMLR, 2021.
  30. V. Bychkovsky, S. Paris, E. Chan, and F. Durand, “Learning photographic global tonal adjustment with a database of input/output image pairs,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 97–104.   IEEE, 2011.
  31. S. Nah, T. Hyun Kim, and K. Mu Lee, “Deep multi-scale convolutional neural network for dynamic scene deblurring,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3883–3891, 2017.
  32. Z. Shen, W. Wang, X. Lu, J. Shen, H. Ling, T. Xu, and L. Shao, “Human-aware motion deblurring,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 5572–5581, 2019.
  33. S. Su, M. Delbracio, J. Wang, G. Sapiro, W. Heidrich, and O. Wang, “Deep video deblurring for hand-held cameras,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1279–1288, 2017.
  34. H. Zhang and V. M. Patel, “Density-aware single image de-raining using a multi-stream dense network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 695–704, 2018.
  35. R. Timofte, S. Gu, J. Wu, and L. Van Gool, “Ntire 2018 challenge on single image super-resolution: Methods and results,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 852–863, 2018.
  36. E. Agustsson and R. Timofte, “Ntire 2017 challenge on single image super-resolution: Dataset and study,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 126–135, 2017.
  37. M. Arndt, O. Nairz, J. Vos-Andreae, C. Keller, G. Van der Zouw, and A. Zeilinger, “Wave–particle duality of c60 molecules,” Nature, vol. 401, no. 6754, pp. 680–682, 1999.
  38. E. Heller, M. Crommie, C. Lutz, and D. Eigler, “Scattering and absorption of surface electron waves in quantum corrals,” Nature, vol. 369, no. 6480, pp. 464–466, 1994.
  39. P. Arbelaez, M. Maire, C. Fowlkes, and J. Malik, “Contour detection and hierarchical image segmentation,” IEEE transactions on pattern analysis and machine intelligence, vol. 33, no. 5, pp. 898–916, 2010.
  40. S. Gu, Y. Li, L. V. Gool, and R. Timofte, “Self-guided network for fast image denoising,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 2511–2520, 2019.
  41. S. Roth and M. J. Black, “Fields of experts,” International Journal of Computer Vision, vol. 82, no. 2, pp. 205–229, 2009.
  42. Huawei, “Mindspore,” 2020. [Online]. Available: https://www.mindspore.cn/
  43. X. Glorot and Y. Bengio, “Understanding the difficulty of training deep feedforward neural networks,” in Proceedings of the thirteenth international conference on artificial intelligence and statistics, pp. 249–256.   JMLR Workshop and Conference Proceedings, 2010.
  44. K. He, X. Zhang, S. Ren, and J. Sun, “Delving deep into rectifiers: Surpassing human-level performance on imagenet classification,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1026–1034, 2015.
  45. D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
  46. B. Sun and K. Saenko, “Deep coral: Correlation alignment for deep domain adaptation,” in European Conference on Computer Vision, pp. 443–450.   Springer, 2016.
  47. S. Cui, S. Wang, J. Zhuo, L. Li, Q. Huang, and Q. Tian, “Towards discriminability and diversity: Batch nuclear-norm maximization under label insufficient situations,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3941–3950, 2020.
  48. S. Anwar and N. Barnes, “Real image denoising with feature attention,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3155–3164, 2019.
  49. S. W. Zamir, A. Arora, S. Khan, M. Hayat, F. S. Khan, M.-H. Yang, and L. Shao, “Multi-stage progressive image restoration,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 14 821–14 831, 2021.
  50. S. W. Zamir, A. Arora, S. Khan, M. Hayat, F. S. Khan, M.-H. Yang, and L. Shao, “Learning enriched features for real image restoration and enhancement,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXV 16, pp. 492–511.   Springer, 2020.
  51. R. Yasarla, V. A. Sindagi, and V. M. Patel, “Syn2real transfer learning for image deraining using gaussian processes,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2726–2736, 2020.
  52. R. Neshatavar, M. Yavartanoo, S. Son, and K. M. Lee, “Cvf-sid: Cyclic multi-variate function for self-supervised image denoising by disentangling noise from image,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 17 583–17 591, 2022.
  53. Y. Yang, C. Wang, R. Liu, L. Zhang, X. Guo, and D. Tao, “Self-augmented unpaired image dehazing via density and depth decomposition,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2037–2046, 2022.
  54. O. Özdenizci and R. Legenstein, “Restoring vision in adverse weather conditions with patch-based denoising diffusion models,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  55. B. Cai, X. Xu, K. Jia, C. Qing, and D. Tao, “Dehazenet: An end-to-end system for single image haze removal,” IEEE Transactions on Image Processing, vol. 25, no. 11, pp. 5187–5198, 2016.
  56. W. Ren, L. Ma, J. Zhang, J. Pan, X. Cao, W. Liu, and M.-H. Yang, “Gated fusion network for single image dehazing,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3253–3261, 2018.
  57. D. Engin, A. Genç, and H. Kemal Ekenel, “Cycle-dehaze: Enhanced cyclegan for single image dehazing,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 825–833, 2018.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Authors (2)