Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
93 tokens/sec
Gemini 2.5 Pro Premium
54 tokens/sec
GPT-5 Medium
22 tokens/sec
GPT-5 High Premium
17 tokens/sec
GPT-4o
101 tokens/sec
DeepSeek R1 via Azure Premium
91 tokens/sec
GPT OSS 120B via Groq Premium
441 tokens/sec
Kimi K2 via Groq Premium
225 tokens/sec
2000 character limit reached

GABIC: Graph-based Attention Block for Image Compression (2410.02981v1)

Published 3 Oct 2024 in eess.IV, cs.CV, and cs.LG

Abstract: While standardized codecs like JPEG and HEVC-intra represent the industry standard in image compression, neural Learned Image Compression (LIC) codecs represent a promising alternative. In detail, integrating attention mechanisms from Vision Transformers into LIC models has shown improved compression efficiency. However, extra efficiency often comes at the cost of aggregating redundant features. This work proposes a Graph-based Attention Block for Image Compression (GABIC), a method to reduce feature redundancy based on a k-Nearest Neighbors enhanced attention mechanism. Our experiments show that GABIC outperforms comparable methods, particularly at high bit rates, enhancing compression performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. Gregory K Wallace, “The JPEG still picture compression standard,” CACM, 1991.
  2. Bellard Fabrice, “BPG image format,” 2014.
  3. Benjamin Bross et al., “Overview of the versatile video coding (VVC) standard and its applications,” IEEE TCSVT, 2021.
  4. Johannes Ballé et al., “End-to-end optimized image compression,” in ICLR, 2017.
  5. Johannes Ballé et al., “Variational image compression with a scale hyperprior,” in ICLR, 2018.
  6. “Joint autoregressive and hierarchical priors for learned image compression,” NeurIPS, 2018.
  7. Zhengxue Cheng et al., “Learned image compression with discretized gaussian mixture likelihoods and attention modules,” in CVPR, 2020.
  8. “The devil is in the details: Window-based attention for image compression,” in CVPR, 2022.
  9. “Density modeling of images using a generalized normalization transformation,” in ICLR, 2016.
  10. “Learned image compression with mixed transformer-CNN architectures,” in CVPR, 2023.
  11. Ze Liu et al., “Swin transformer: Hierarchical vision transformer using shifted windows,” in ICCV, 2021.
  12. Petar Veličković et al., “Graph attention networks,” in ICLR, 2018.
  13. Siwei Ma et al., “Image and video compression with neural networks: A review,” IEEE TCSVT, 2019.
  14. Alberto Presta et al., “A differentiable entropy model for learned image compression,” in ICIAP, 2023.
  15. “Channel-wise autoregressive entropy models for learned image compression,” in ICIP, 2020.
  16. “Enhanced invertible encoding for learned image compression,” in ACMMM, 2021.
  17. Chunhui Yang et al., “Graph-convolution network for image compression,” in ICIP, 2021.
  18. Lei Zhou et al., “End-to-end optimized image compression with attention mechanism,” in CVPR Workshops, 2019.
  19. Haojie Liu et al., “Non-local attention optimized deep image compression,” ArXiv, 2019.
  20. “Semi-supervised classification with graph convolutional networks,” in ICLR, 2017.
  21. “Vision GNN: An image is worth graph of nodes,” NeurIPS, 2022.
  22. Yan Han et al., “Vision HGNN: An image is more than a graph of nodes,” in ICCV, 2023.
  23. Jhony H. Giraldo et al., “Hypergraph convolutional networks for weakly-supervised semantic segmentation,” in ICIP, 2022.
  24. Alexey Dosovitskiy et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” in ICLR, 2021.
  25. Justin Gilmer et al., “Neural message passing for quantum chemistry,” in ICML, 2017.
  26. Xiaolong Wang et al., “Non-local neural networks,” in CVPR, 2018.
  27. Jean Bégaint et al., “CompressAI: a PyTorch library and evaluation platform for end-to-end compression research,” arXiv, 2020.
  28. “Fast graph representation learning with PyTorch Geometric,” in ICLR Workshops, 2019.
  29. Ivan Krasin et al., “OpenImages: A public dataset for large-scale multi-label and multi-class image classification.,” Dataset available from https://github.com/openimages, 2016.
  30. “Adam: A method for stochastic optimization,” in ICLR, 2015.
  31. Eastman Kodak, “Kodak lossless true color image suite (photocd pcd0992),” 1993, http://r0k.us/graphics/kodak/.
  32. “Workshop and challenge on learned image compression,” 2020, https://www.compression.cc/.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube