Taming Lookup Tables for Efficient Image Retouching (2403.19238v2)
Abstract: The widespread use of high-definition screens in edge devices, such as end-user cameras, smartphones, and televisions, is spurring a significant demand for image enhancement. Existing enhancement models often optimize for high performance while falling short of reducing hardware inference time and power consumption, especially on edge devices with constrained computing and storage resources. To this end, we propose Image Color Enhancement Lookup Table (ICELUT) that adopts LUTs for extremely efficient edge inference, without any convolutional neural network (CNN). During training, we leverage pointwise (1x1) convolution to extract color information, alongside a split fully connected layer to incorporate global information. Both components are then seamlessly converted into LUTs for hardware-agnostic deployment. ICELUT achieves near-state-of-the-art performance and remarkably low power consumption. We observe that the pointwise network structure exhibits robust scalability, upkeeping the performance even with a heavily downsampled 32x32 input image. These enable ICELUT, the first-ever purely LUT-based image enhancer, to reach an unprecedented speed of 0.4ms on GPU and 7ms on CPU, at least one order faster than any CNN solution. Codes are available at https://github.com/Stephen0808/ICELUT.
- Y. Jo and S. J. Kim, “Practical single-image super-resolution using look-up table,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 691–700.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- W. Ouyang, Y. Dong, X. Kang, P. Ren, X. Xu, and X. Xie, “Rsfnet: A white-box image retouching approach using region-specific color filters,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 12 160–12 169.
- C. Liu, H. Yang, J. Fu, and X. Qian, “4d lut: learnable context-aware 4d lookup table for image enhancement,” IEEE Transactions on Image Processing, vol. 32, pp. 4742–4756, 2023.
- J. Li, C. Chen, Z. Cheng, and Z. Xiong, “Mulut: Cooperating multiple look-up tables for efficient image super-resolution,” in European Conference on Computer Vision. Springer, 2022, pp. 238–256.
- F. Zhang, H. Zeng, T. Zhang, and L. Zhang, “Clut-net: Learning adaptively compressed representations of 3dluts for lightweight image enhancement,” in Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 6493–6501.
- C. Ma, J. Zhang, J. Zhou, and J. Lu, “Learning series-parallel lookup tables for efficient image super-resolution,” in European Conference on Computer Vision. Springer, 2022, pp. 305–321.
- H. Zeng, J. Cai, L. Li, Z. Cao, and L. Zhang, “Learning image-adaptive 3d lookup tables for high performance photo enhancement in real-time,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 4, pp. 2058–2073, 2020.
- V. Bychkovsky, S. Paris, E. Chan, and F. Durand, “Learning photographic global tonal adjustment with a database of input/output image pairs,” in CVPR 2011. IEEE, 2011, pp. 97–104.
- M. Afifi, B. Price, S. Cohen, and M. S. Brown, “When color constancy goes wrong: Correcting improperly white-balanced images,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 1535–1544.
- H.-U. Kim, Y. J. Koh, and C.-S. Kim, “Global and local enhancement networks for paired and unpaired image enhancement,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXV 16. Springer, 2020, pp. 339–354.
- S. Moran, S. McDonagh, and G. Slabaugh, “Curl: Neural curve layers for global image enhancement,” in 2020 25th International Conference on Pattern Recognition (ICPR). IEEE, 2021, pp. 9796–9803.
- M. Gharbi, J. Chen, J. T. Barron, S. W. Hasinoff, and F. Durand, “Deep bilateral learning for real-time image enhancement,” ACM Transactions on Graphics (TOG), vol. 36, no. 4, pp. 1–12, 2017.
- S. Moran, P. Marza, S. McDonagh, S. Parisot, and G. Slabaugh, “Deeplpf: Deep local parametric filters for image enhancement,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 12 826–12 835.
- J. He, Y. Liu, Y. Qiao, and C. Dong, “Conditional sequential modulation for efficient global image retouching,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XIII 16. Springer, 2020, pp. 679–695.
- M. Afifi and M. S. Brown, “Deep white-balance editing,” in Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition, 2020, pp. 1397–1406.
- Y.-S. Chen, Y.-C. Wang, M.-H. Kao, and Y.-Y. Chuang, “Deep photo enhancer: Unpaired learning for image enhancement from photographs with gans,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 6306–6314.
- H.-U. Kim, Y. J. Koh, and C.-S. Kim, “Pienet: Personalized image enhancement network,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXX 16. Springer, 2020, pp. 374–390.
- R. Wang, Q. Zhang, C.-W. Fu, X. Shen, W.-S. Zheng, and J. Jia, “Underexposed photo enhancement using deep illumination estimation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 6849–6857.
- Y. Shih, S. Paris, F. Durand, and W. T. Freeman, “Data-driven hallucination of different times of day from a single outdoor photo,” ACM Transactions on Graphics (TOG), vol. 32, no. 6, pp. 1–11, 2013.
- T. Wang, Y. Li, J. Peng, Y. Ma, X. Wang, F. Song, and Y. Yan, “Real-time image enhancer via learnable spatial-aware 3d lookup tables,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 2471–2480.
- C. Yang, M. Jin, Y. Xu, R. Zhang, Y. Chen, and H. Liu, “Seplut: Separable image-adaptive lookup tables for real-time image enhancement,” in European Conference on Computer Vision. Springer, 2022, pp. 201–217.
- J. Liang, H. Zeng, M. Cui, X. Xie, and L. Zhang, “Ppr10k: A large-scale portrait photo retouching dataset with human-region mask and group-level consistency,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 653–661.
- Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE transactions on image processing, vol. 13, no. 4, pp. 600–612, 2004.
- B. Huang, J. C. L. Li, J. Ran, B. Li, J. Zhou, D. Yu, and N. Wong, “Hundred-kilobyte lookup tables for efficient single-image super-resolution,” arXiv preprint arXiv:2312.06101, 2023.