Exploration of Learned Lifting-Based Transform Structures for Fully Scalable and Accessible Wavelet-Like Image Compression (2402.18761v1)
Abstract: This paper provides a comprehensive study on features and performance of different ways to incorporate neural networks into lifting-based wavelet-like transforms, within the context of fully scalable and accessible image compression. Specifically, we explore different arrangements of lifting steps, as well as various network architectures for learned lifting operators. Moreover, we examine the impact of the number of learned lifting steps, the number of channels, the number of layers and the support of kernels in each learned lifting operator. To facilitate the study, we investigate two generic training methodologies that are simultaneously appropriate to a wide variety of lifting structures considered. Experimental results ultimately suggest that retaining fixed lifting steps from the base wavelet transform is highly beneficial. Moreover, we demonstrate that employing more learned lifting steps and more layers in each learned lifting operator do not contribute strongly to the compression performance. However, benefits can be obtained by utilizing more channels in each learned lifting operator. Ultimately, the learned wavelet-like transform proposed in this paper achieves over 25% bit-rate savings compared to JPEG 2000 with compact spatial support.
- “End-to-end optimized image compression,” arXiv preprint arXiv:1611.01704, 2016.
- “Joint autoregressive and hierarchical priors for learned image compression,” Advances in neural information processing systems, vol. 31, 2018.
- “Lossy image compression with compressive autoencoders,” arXiv preprint arXiv:1703.00395, 2017.
- “Full resolution image compression with recurrent neural networks,” in Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2017, pp. 5306–5314.
- “iwave: Cnn-based wavelet-like transform for image compression,” IEEE Transactions on Multimedia, vol. 22, no. 7, pp. 1667–1679, 2019.
- “Optimized lifting scheme based on a dynamical fully connected network for image coding,” in 2020 IEEE International Conference on Image Processing (ICIP). IEEE, 2020, pp. 3329–3333.
- “A neural network approach for joint optimization of predictors in lifting-based image coders,” in 2021 IEEE International Conference on Image Processing (ICIP). IEEE, 2021, pp. 3747–3751.
- “Dynamic neural network for lossy-to-lossless image coding,” IEEE Transactions on Image Processing, vol. 31, pp. 569–584, 2021.
- “Machine-learning based secondary transform for improved image compression in jpeg2000,” in 2021 IEEE International Conference on Image Processing (ICIP), 2021, pp. 3752–3756.
- “A neural network lifting based secondary transform for improved fully scalable image compression in jpeg 2000,” in 2022 IEEE International Conference on Image Processing (ICIP), 2022, pp. 1606–1610.
- “Neural network assisted lifting steps for improved fully scalable lossy image compression in jpeg2000,” IEEE Transactions on Pattern Analysis and Machine Intelligence, to be published.
- Multiresolution signal decomposition: transforms, subbands, and wavelets, Academic press, 2001.
- Wim Sweldens, “The lifting scheme: A construction of second generation wavelets,” SIAM journal on mathematical analysis, vol. 29, no. 2, pp. 511–546, 1998.
- Wim Sweldens, “Lifting scheme: a new philosophy in biorthogonal wavelet constructions,” in Wavelet Applications in Signal and Image Processing III, Andrew F. Laine and Michael A. Unser, Eds. International Society for Optics and Photonics, 1995, vol. 2569, pp. 68 – 79, SPIE.
- Wim Sweldens, “The lifting scheme: A custom-design construction of biorthogonal wavelets,” Applied and Computational Harmonic Analysis, vol. 3, no. 2, pp. 186–200, 1996.
- “Factoring wavelet transforms into lifting steps,” Journal of Fourier analysis and applications, vol. 4, no. 3, pp. 247–269, 1998.
- “Factoring wavelet transforms into lifting steps,” Wavelets in the Geosciences, pp. 131–157, 2005.
- “U-net: Convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Springer, 2015, pp. 234–241.
- “Block-optimized variable bit rate neural image compression,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2018, pp. 2551–2554.
- “Learning-based image compression using convolutional autoencoder and wavelet decomposition,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2019, number CONF.
- “Variational image compression with a scale hyperprior,” arXiv preprint arXiv:1802.01436, 2018.
- “Coarse-to-fine hyper-prior modeling for learned image compression,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2020, vol. 34, pp. 11013–11020.
- “Rate-distortion optimized learning-based image compression using an adaptive hierachical autoencoder with conditional hyperprior,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 1885–1889.
- “A spatial rnn codec for end-to-end image compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 13269–13277.
- “Context-adaptive entropy model for end-to-end optimized image compression,” arXiv preprint arXiv:1809.10452, 2018.
- “Efficient and effective context-based convolutional entropy modeling for image compression,” IEEE Transactions on Image Processing, vol. 29, pp. 5900–5911, 2020.
- “Checkerboard context model for efficient learned image compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 14771–14780.
- “Elic: Efficient learned image compression with unevenly grouped space-channel contextual adaptive coding,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5718–5727.
- “Biorthogonal bases of compactly supported wavelets,” Communications on pure and applied mathematics, vol. 45, no. 5, pp. 485–560, 1992.
- JPEG 2000: Image Compression Fundamentals, Standards and Practice, Kluwer Academic Publishers, Norwell, MA, USA, 2001.
- “End-to-end optimized versatile image compression with wavelet-like transform,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 3, pp. 1247–1263, 2020.
- “Improved transform structures for learned wavelet-like fully scalable image compression,” 2023 IEEE 25th International Workshop on Multimedia Signal Processing (MMSP), 2023.
- “Nonlinear wavelet transforms for image coding via lifting,” IEEE Transactions on Image Processing, vol. 12, no. 12, pp. 1449–1459, 2003.
- “Learning convolutional networks for content-weighted image compression,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3214–3223.
- “Universally quantized neural compression,” Advances in neural information processing systems, vol. 33, pp. 12367–12376, 2020.
- “Estimating or propagating gradients through stochastic neurons for conditional computation,” arXiv preprint arXiv:1308.3432, 2013.
- “Understanding straight-through estimator in training activation quantized neural nets,” arXiv preprint arXiv:1903.05662, 2019.
- “Soft-to-hard vector quantization for end-to-end learning compressible representations,” Advances in neural information processing systems, vol. 30, 2017.
- “Improving inference for neural image compression,” Advances in Neural Information Processing Systems, vol. 33, pp. 573–584, 2020.
- Sepp Hochreiter, “The vanishing gradient problem during learning recurrent neural nets and problem solutions,” International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems, vol. 6, no. 02, pp. 107–116, 1998.
- “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
- Gisle Bjontegaard, “Calculation of average psnr differences between rd-curves,” ITU SG16 Doc. VCEG-M33, 2001.