Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Flexible Variable-Rate Image Feature Compression for Edge-Cloud Systems (2404.00432v1)

Published 30 Mar 2024 in eess.IV

Abstract: Feature compression is a promising direction for coding for machines. Existing methods have made substantial progress, but they require designing and training separate neural network models to meet different specifications of compression rate, performance accuracy and computational complexity. In this paper, a flexible variable-rate feature compression method is presented that can operate on a range of rates by introducing a rate control parameter as an input to the neural network model. By compressing different intermediate features of a pre-trained vision task model, the proposed method can scale the encoding complexity without changing the overall size of the model. The proposed method is more flexible than existing baselines, at the same time outperforming them in terms of the three-way trade-off between feature compression rate, vision task accuracy, and encoding complexity. We have made the source code available at https://github.com/adnan-hossain/var_feat_comp.git.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. “Searching for mobilenetv3,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 1314–1324.
  2. “Mobilenetv2: Inverted residuals and linear bottlenecks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 4510–4520.
  3. “Jpeg2000: Image compression fundamentals, standards and practice,” Journal of Electronic Imaging, vol. 11, no. 2, pp. 286–287, 2002.
  4. “Intra coding of the hevc standard,” IEEE transactions on circuits and systems for video technology, vol. 22, no. 12, pp. 1792–1801, 2012.
  5. “End-to-end optimized image compression,” arXiv preprint arXiv:1611.01704, 2016.
  6. “Soft-to-hard vector quantization for end-to-end learning compressible representations,” Advances in neural information processing systems, vol. 30, 2017.
  7. “Variational image compression with a scale hyperprior,” arXiv preprint arXiv:1802.01436, 2018.
  8. “Context-adaptive entropy model for end-to-end optimized image compression,” arXiv preprint arXiv:1809.10452, 2018.
  9. “Efficient feature compression for edge-cloud systems,” in 2022 Picture Coding Symposium (PCS). IEEE, 2022, pp. 187–191.
  10. “End-to-end learning of compressible features,” in 2020 IEEE International Conference on Image Processing (ICIP). IEEE, 2020, pp. 3349–3353.
  11. “Lightweight compression of neural network feature tensors for collaborative intelligence,” in IEEE International Conference on Multimedia and Expo, July 2020, pp. 1–6.
  12. “Supervised compression for resource-constrained edge computing systems,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2022, pp. 2685–2695.
  13. “A low-complexity approach to rate-distortion optimized variable bit-rate compression for split dnn computing,” in 2022 26th International Conference on Pattern Recognition (ICPR). IEEE, 2022, pp. 182–188.
  14. “Variable rate deep image compression with a conditional autoencoder,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 3146–3154.
  15. “Lightweight compression of intermediate neural network features for collaborative intelligence,” IEEE Open Journal of Circuits and Systems, vol. 2, pp. 350–362, May 2021.
  16. Jarek Duda, “Asymmetric numeral systems: entropy coding combining speed of huffman coding with compression rate of arithmetic coding,” arXiv preprint arXiv:1311.2540, 2013.
  17. “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  18. “Scalable image coding for humans and machines,” IEEE Transactions on Image Processing, vol. 31, pp. 2739–2754, 2022.
  19. “A theory of usable information under computational constraints,” arXiv preprint arXiv:2002.10689, 2020.
  20. “A convnet for the 2020s,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 11976–11986.
  21. “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2818–2826.
  22. “Qarv: Quantization-aware resnet vae for lossy image compression,” arXiv preprint arXiv:2302.08899, 2023.
  23. “Conditional image generation with pixelcnn decoders,” Advances in neural information processing systems, vol. 29, 2016.
  24. “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
  25. “Neural image compression for gigapixel histopathology image analysis,” IEEE transactions on pattern analysis and machine intelligence, vol. 43, no. 2, pp. 567–578, 2019.
  26. “Imagenet: A large-scale hierarchical image database,” in 2009 IEEE conference on computer vision and pattern recognition. Ieee, 2009, pp. 248–255.
  27. Gisle Bjontegaard, “Calculation of average psnr differences between rd-curves,” ITU SG16 Doc. VCEG-M33, 2001.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Md Adnan Faisal Hossain (3 papers)
  2. Zhihao Duan (38 papers)
  3. Yuning Huang (11 papers)
  4. Fengqing Zhu (77 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.