ProgDTD: Progressive Learned Image Compression with Double-Tail-Drop Training (2305.02145v2)
Abstract: Progressive compression allows images to start loading as low-resolution versions, becoming clearer as more data is received. This increases user experience when, for example, network connections are slow. Today, most approaches for image compression, both classical and learned ones, are designed to be non-progressive. This paper introduces ProgDTD, a training method that transforms learned, non-progressive image compression approaches into progressive ones. The design of ProgDTD is based on the observation that the information stored within the bottleneck of a compression model commonly varies in importance. To create a progressive compression model, ProgDTD modifies the training steps to enforce the model to store the data in the bottleneck sorted by priority. We achieve progressive compression by transmitting the data in order of its sorted index. ProgDTD is designed for CNN-based learned image compression models, does not need additional parameters, and has a customizable range of progressiveness. For evaluation, we apply ProgDTDto the hyperprior model, one of the most common structures in learned image compression. Our experimental results show that ProgDTD performs comparably to its non-progressive counterparts and other state-of-the-art progressive models in terms of MS-SSIM and accuracy.
- Generative adversarial networks for extreme learned image compression. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 221–231, 2019.
- Density modeling of images using a generalized normalization transformation. arXiv preprint arXiv:1511.06281, 2015.
- End-to-end optimized image compression. arXiv preprint arXiv:1611.01704, 2016.
- Variational image compression with a scale hyperprior. arXiv preprint arXiv:1802.01436, 2018.
- Bpg image format. https://bellard.org/bpg/. Accessed: 2023-02-14.
- Efficient variable rate image compression with multi-scale decomposition network. IEEE Transactions on Circuits and Systems for Video Technology, 29(12):3687–3700, 2018.
- A novel deep progressive image compression framework. In 2019 Picture Coding Symposium (PCS), pages 1–5. IEEE, 2019.
- Learned image compression with discretized gaussian mixture likelihoods and attention modules. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7939–7948, 2020.
- Variable rate deep image compression with a conditional autoencoder. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3146–3154, 2019.
- Asymmetric gained deep image compression with continuous rate adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10532–10541, 2021.
- Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
- Drasic: Distributed recurrent autoencoder for scalable image compression. In 2020 Data Compression Conference (DCC), pages 3–12. IEEE, 2020.
- Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020.
- Towards conceptual compression. Advances In Neural Information Processing Systems, 29, 2016.
- Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
- Improved lossy image compression with priming and spatially adaptive bit rates for recurrent networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4385–4393, 2018.
- Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Edge-host partitioning of deep neural networks with feature space encoding for resource-constrained internet-of-things platforms. In 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pages 1–6. IEEE, 2018.
- Eastman kodak (1993). kodak lossless true color image suite (photocd pcd0992). https://r0k.us/graphics/kodak.
- Stochastic bottleneck: Rateless auto-encoder for flexible dimensionality reduction. In 2020 IEEE International Symposium on Information Theory (ISIT), pages 2735–2740. IEEE, 2020.
- Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60(6):84–90, 2017.
- Compression of black-white images with arithmetic coding. IEEE Transactions on Communications, 29(6):858–867, 1981.
- Context-adaptive entropy model for end-to-end optimized image compression. arXiv preprint arXiv:1809.10452, 2018.
- Dpict: Deep progressive image compression using trit-planes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16113–16122, 2022.
- Jalad: Joint accuracy-and latency-aware deep structure decoupling for edge-cloud execution. In 2018 IEEE 24th international conference on parallel and distributed systems (ICPADS), pages 671–678. IEEE, 2018.
- Joint autoregressive and hierarchical priors for learned image compression. Advances in neural information processing systems, 31, 2018.
- Neuralcompression. https://github.com/facebookresearch/NeuralCompression, 2021.
- Neural multi-scale image compression. In Computer Vision–ACCV 2018: 14th Asian Conference on Computer Vision, Perth, Australia, December 2–6, 2018, Revised Selected Papers, Part VI 14, pages 718–732. Springer, 2019.
- J-R Ohm. Advances in scalable video coding. Proceedings of the IEEE, 93(1):42–56, 2005.
- Real-time adaptive image compression. In International Conference on Machine Learning, pages 2922–2930. PMLR, 2017.
- Nonlinear principal component analysis: neural network models and applications. In Principal manifolds for data visualization and dimension reduction, pages 44–67. Springer, 2008.
- The jpeg 2000 still image compression standard. IEEE Signal processing magazine, 18(5):36–58, 2001.
- Scalable learned image compression with a recurrent neural networks-based hyperprior. In 2020 IEEE International Conference on Image Processing (ICIP), pages 3369–3373. IEEE, 2020.
- Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning, pages 6105–6114. PMLR, 2019.
- Lossy image compression with compressive autoencoders. arXiv preprint arXiv:1703.00395, 2017.
- Variable rate image compression with recurrent neural networks. arXiv preprint arXiv:1511.06085, 2015.
- Full resolution image compression with recurrent neural networks. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pages 5306–5314, 2017.
- Deep generative models for distribution-preserving lossy compression. Advances in neural information processing systems, 31, 2018.
- Pixel recurrent neural networks. In International conference on machine learning, pages 1747–1756. PMLR, 2016.
- Gregory K Wallace. The jpeg still picture compression standard. Communications of the ACM, 34(4):30–44, 1991.
- Gregory K Wallace. The jpeg still picture compression standard. IEEE transactions on consumer electronics, 38(1):xviii–xxxiv, 1992.
- Webp. https://developers.google.com/speed/webp/docs/compression. Accessed: 2023-02-14.
- Principal component analysis. Chemometrics and intelligent laboratory systems, 2(1-3):37–52, 1987.
- Video enhancement with task-oriented flow. International Journal of Computer Vision, 127:1106–1125, 2019.
- Slimmable compressive autoencoders for practical neural image compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4998–5007, 2021.
- Variable rate deep image compression with modulated autoencoder. IEEE Signal Processing Letters, 27:331–335, 2020.
- Deep compressive offloading: Speeding up neural network inference by trading edge computation for network latency. In Proceedings of the 18th Conference on Embedded Networked Sensor Systems, pages 476–488, 2020.
- Fastdeepiot: Towards understanding and optimizing neural network execution time on mobile and embedded devices. In Proceedings of the 16th ACM Conference on Embedded Networked Sensor Systems, pages 278–291, 2018.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.