FLLIC: Functionally Lossless Image Compression (2401.13616v2)
Abstract: Recently, DNN models for lossless image coding have surpassed their traditional counterparts in compression performance, reducing the bit rate by about ten percent for natural color images. But even with these advances, mathematically lossless image compression (MLLIC) ratios for natural images still fall short of the bandwidth and cost-effectiveness requirements of most practical imaging and vision systems at present and beyond. To break the bottleneck of MLLIC in compression performance, we question the necessity of MLLIC, as almost all digital sensors inherently introduce acquisition noises, making mathematically lossless compression counterproductive. Therefore, in contrast to MLLIC, we propose a new paradigm of joint denoising and compression called functionally lossless image compression (FLLIC), which performs lossless compression of optimally denoised images (the optimality may be task-specific). Although not literally lossless with respect to the noisy input, FLLIC aims to achieve the best possible reconstruction of the latent noise-free original image. Extensive experiments show that FLLIC achieves state-of-the-art performance in joint denoising and compression of noisy images and does so at a lower computational cost.
- J. Ballé, V. Laparra, and E. P. Simoncelli, “End-to-end optimized image compression,” in 5th International Conference on Learning Representations, ICLR, 2017.
- L. Theis, W. Shi, A. Cunningham, and F. Huszár, “Lossy image compression with compressive autoencoders,” in 5th International Conference on Learning Representations, ICLR, 2017.
- E. Agustsson, F. Mentzer, M. Tschannen, L. Cavigelli, R. Timofte, L. Benini, and L. V. Gool, “Soft-to-hard vector quantization for end-to-end learning compressible representations,” in Advances in Neural Information Processing Systems 30, 2017, pp. 1141–1151.
- J. Ballé, D. Minnen, S. Singh, S. J. Hwang, and N. Johnston, “Variational image compression with a scale hyperprior,” in 6th International Conference on Learning Representations, ICLR. OpenReview.net, 2018.
- D. Minnen, J. Ballé, and G. Toderici, “Joint autoregressive and hierarchical priors for learned image compression,” in Advances in Neural Information Processing Systems 31, 2018, pp. 10 794–10 803.
- F. Mentzer, E. Agustsson, M. Tschannen, R. Timofte, and L. V. Gool, “Conditional probability models for deep image compression,” in 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2018, pp. 4394–4402.
- J. Lee, S. Cho, and S. Beack, “Context-adaptive entropy model for end-to-end optimized image compression,” in 7th International Conference on Learning Representations, ICLR, 2019.
- Z. Cheng, H. Sun, M. Takeuchi, and J. Katto, “Learned image compression with discretized gaussian mixture likelihoods and attention modules,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR, 2020, pp. 7936–7945.
- F. Mentzer, G. D. Toderici, M. Tschannen, and E. Agustsson, “High-fidelity generative image compression,” Advances in Neural Information Processing Systems, vol. 33, pp. 11 913–11 924, 2020.
- X. Zhang and X. Wu, “Attention-guided image compression by deep reconstruction of compressive sensed saliency skeleton,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 13 354–13 364.
- D. He, Y. Zheng, B. Sun, Y. Wang, and H. Qin, “Checkerboard context model for efficient learned image compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2021, pp. 14 771–14 780.
- F. Yang, L. Herranz, Y. Cheng, and M. G. Mozerov, “Slimmable compressive autoencoders for practical neural image compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2021, pp. 4998–5007.
- J.-H. Kim, B. Heo, and J.-S. Lee, “Joint global and local hierarchical priors for learned image compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5992–6001.
- D. He, Z. Yang, W. Peng, R. Ma, H. Qin, and Y. Wang, “Elic: Efficient learned image compression with unevenly grouped space-channel contextual adaptive coding,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5718–5727.
- C. Gao, T. Xu, D. He, Y. Wang, and H. Qin, “Flexible neural image compression via code editing,” Advances in Neural Information Processing Systems, vol. 35, pp. 12 184–12 196, 2022.
- T. Xu, Y. Wang, D. He, C. Gao, H. Gao, K. Liu, and H. Qin, “Multi-sample training for neural image compression,” arXiv preprint arXiv:2209.13834, 2022.
- J. Lee, S. Jeong, and M. Kim, “Selective compression learning of latent representations for variable-rate image compression,” arXiv preprint arXiv:2211.04104, 2022.
- C. Shin, H. Lee, H. Son, S. Lee, D. Lee, and S. Lee, “Expanded adaptive scaling normalization for end to end image compression,” in European Conference on Computer Vision. Springer, 2022, pp. 390–405.
- X. Zhang and X. Wu, “Dual-layer image compression via adaptive downsampling and spatially varying upconversion,” arXiv preprint arXiv:2302.06096, 2023.
- R. Zou, C. Song, and Z. Zhang, “The devil is in the details: Window-based attention for image compression,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 17 492–17 501.
- X. Zhang and X. Wu, “Lvqac: Lattice vector quantization coupled with spatially adaptive companding for efficient learned image compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 10 239–10 248.
- F. Mentzer, E. Agustsson, M. Tschannen, R. Timofte, and L. V. Gool, “Practical full resolution learned lossless image compression,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 10 629–10 638.
- F. Mentzer, L. V. Gool, and M. Tschannen, “Learning better lossless compression using lossy compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 6638–6647.
- X. Zhang and X. Wu, “Nonlinear prediction of multidimensional signals via deep regression with applications to image coding,” in ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019, pp. 1602–1606.
- F. Kingma, P. Abbeel, and J. Ho, “Bit-swap: Recursive bits-back coding for lossless compression with hierarchical latent variables,” in International Conference on Machine Learning. PMLR, 2019, pp. 3408–3417.
- J. Townsend, T. Bird, J. Kunze, and D. Barber, “Hilloc: Lossless image compression with hierarchical latent variable models,” arXiv preprint arXiv:1912.09953, 2019.
- E. Hoogeboom, J. Peters, R. Van Den Berg, and M. Welling, “Integer discrete flows and lossless compression,” Advances in Neural Information Processing Systems, vol. 32, 2019.
- J. Ho, E. Lohn, and P. Abbeel, “Compression with flows via local bits-back coding,” Advances in Neural Information Processing Systems, vol. 32, 2019.
- X. Zhang and X. Wu, “Ultra high fidelity deep image decompression with ℓ∞subscriptℓ\ell_{\infty}roman_ℓ start_POSTSUBSCRIPT ∞ end_POSTSUBSCRIPT-constrained compression,” IEEE Transactions on Image Processing, vol. 30, pp. 963–975, 2020.
- S. Zhang, N. Kang, T. Ryder, and Z. Li, “iflow: Numerically invertible flows for efficient lossless compression via a uniform coder,” Advances in Neural Information Processing Systems, vol. 34, pp. 5822–5833, 2021.
- S. Zhang, C. Zhang, N. Kang, and Z. Li, “ivpf: Numerical invertible volume preserving flow for efficient lossless compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 620–629.
- N. Kang, S. Qiu, S. Zhang, Z. Li, and S.-T. Xia, “Pilc: Practical image lossless compression with an end-to-end gpu oriented neural framework,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 3739–3748.
- A. Van Den Oord, N. Kalchbrenner, and K. Kavukcuoglu, “Pixel recurrent neural networks,” in International conference on machine learning. PMLR, 2016, pp. 1747–1756.
- T. Salimans, A. Karpathy, X. Chen, and D. P. Kingma, “Pixelcnn++: Improving the pixelcnn with discretized logistic mixture likelihood and other modifications,” arXiv preprint arXiv:1701.05517, 2017.
- D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” arXiv preprint arXiv:1312.6114, 2013.
- I. Kobyzev, S. J. Prince, and M. A. Brubaker, “Normalizing flows: An introduction and review of current methods,” IEEE transactions on pattern analysis and machine intelligence, vol. 43, no. 11, pp. 3964–3979, 2020.
- X. Zhang, X. Wu, X. Zhai, X. Ben, and C. Tu, “Davd-net: Deep audio-aided video decompression of talking heads,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 12 335–12 344.
- X. Zhang and X. Wu, “Multi-modality deep restoration of extremely compressed face videos,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 2, pp. 2024–2037, 2022.
- K. Zhang, W. Zuo, Y. Chen, D. Meng, and L. Zhang, “Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising,” IEEE transactions on image processing, vol. 26, no. 7, pp. 3142–3155, 2017.
- D. Ulyanov, A. Vedaldi, and V. Lempitsky, “Deep image prior,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 9446–9454.
- T. Huang, S. Li, X. Jia, H. Lu, and J. Liu, “Neighbor2neighbor: Self-supervised denoising from single noisy images,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 14 781–14 790.
- J. Liang, J. Cao, G. Sun, K. Zhang, L. Van Gool, and R. Timofte, “Swinir: Image restoration using swin transformer,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 1833–1844.
- Y. Zhao, Z. Jiang, A. Men, and G. Ju, “Pyramid real image denoising network,” in 2019 IEEE Visual Communications and Image Processing (VCIP). IEEE, 2019, pp. 1–4.
- S. Guo, Z. Yan, K. Zhang, W. Zuo, and L. Zhang, “Toward convolutional blind denoising of real photographs,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 1712–1722.
- C. Ren, X. He, C. Wang, and Z. Zhao, “Adaptive consistency prior based deep network for image denoising,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 8596–8606.
- S. W. Zamir, A. Arora, S. Khan, M. Hayat, F. S. Khan, and M.-H. Yang, “Restormer: Efficient transformer for high-resolution image restoration,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 5728–5739.
- M. Testolina, E. Upenik, and T. Ebrahimi, “Towards image denoising in the latent space of learning-based compression,” in Applications of Digital Image Processing XLIV, vol. 11842. SPIE, 2021, pp. 412–422.
- S. Ranjbar Alvar, M. Ulhaq, H. Choi, and I. V. Bajić, “Joint image compression and denoising via latent-space scalability,” Frontiers in Signal Processing, vol. 2, p. 932873, 2022.
- K. L. Cheng, Y. Xie, and Q. Chen, “Optimizing image compression via joint learning with denoising,” in European Conference on Computer Vision. Springer, 2022, pp. 56–73.
- Y. Huang, Z. Duan, and F. Zhu, “Narv: An efficient noise-adaptive resnet vae for joint image compression and denoising,” in 2023 IEEE International Conference on Multimedia and Expo Workshops (ICMEW). IEEE, 2023, pp. 188–193.
- J. Li, B. Li, and Y. Lu, “Hybrid spatial-temporal entropy modelling for neural video compression,” in Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 1503–1511.
- G.-H. Wang, J. Li, B. Li, and Y. Lu, “Evc: Towards real-time neural image compression with mask decay,” arXiv preprint arXiv:2302.05071, 2023.
- Y. Zhu, Y. Yang, and T. Cohen, “Transformer-based transform coding,” in International Conference on Learning Representations, 2022.
- Y. Qian, M. Lin, X. Sun, Z. Tan, and R. Jin, “Entroformer: A transformer-based entropy model for learned image compression,” in International Conference on Learning Representations, 2022.
- H. Rhee, Y. I. Jang, S. Kim, and N. I. Cho, “Lc-fdnet: Learned lossless image compression with frequency decomposition network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 6033–6042.
- B. Lim, S. Son, H. Kim, S. Nah, and K. Mu Lee, “Enhanced deep residual networks for single image super-resolution,” in Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 2017, pp. 136–144.
- D. Martin, C. Fowlkes, D. Tal, and J. Malik, “A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics,” in Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001, vol. 2. IEEE, 2001, pp. 416–423.
- J.-B. Huang, A. Singh, and N. Ahuja, “Single image super-resolution from transformed self-exemplars,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 5197–5206.
- R. Franzen, “Kodak lossless true color image suite,” 1999, http://r0k.us/graphics/kodak/.
- A. Abdelhamed, S. Lin, and M. S. Brown, “A high-quality denoising dataset for smartphone cameras,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 1692–1700.
- D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
- E. Agustsson and R. Timofte, “Ntire 2017 challenge on single image super-resolution: Dataset and study,” in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, July 2017.