Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 78 tok/s
Gemini 2.5 Pro 43 tok/s Pro
GPT-5 Medium 23 tok/s
GPT-5 High 29 tok/s Pro
GPT-4o 93 tok/s
GPT OSS 120B 470 tok/s Pro
Kimi K2 183 tok/s Pro
2000 character limit reached

Bit-mask Robust Contrastive Knowledge Distillation for Unsupervised Semantic Hashing (2403.06071v1)

Published 10 Mar 2024 in cs.CV and cs.IR

Abstract: Unsupervised semantic hashing has emerged as an indispensable technique for fast image search, which aims to convert images into binary hash codes without relying on labels. Recent advancements in the field demonstrate that employing large-scale backbones (e.g., ViT) in unsupervised semantic hashing models can yield substantial improvements. However, the inference delay has become increasingly difficult to overlook. Knowledge distillation provides a means for practical model compression to alleviate this delay. Nevertheless, the prevailing knowledge distillation approaches are not explicitly designed for semantic hashing. They ignore the unique search paradigm of semantic hashing, the inherent necessities of the distillation process, and the property of hash codes. In this paper, we propose an innovative Bit-mask Robust Contrastive knowledge Distillation (BRCD) method, specifically devised for the distillation of semantic hashing models. To ensure the effectiveness of two kinds of search paradigms in the context of semantic hashing, BRCD first aligns the semantic spaces between the teacher and student models through a contrastive knowledge distillation objective. Additionally, to eliminate noisy augmentations and ensure robust optimization, a cluster-based method within the knowledge distillation process is introduced. Furthermore, through a bit-level analysis, we uncover the presence of redundancy bits resulting from the bit independence property. To mitigate these effects, we introduce a bit mask mechanism in our knowledge distillation objective. Finally, extensive experiments not only showcase the noteworthy performance of our BRCD method in comparison to other knowledge distillation methods but also substantiate the generality of our methods across diverse semantic hashing models and backbones. The code for BRCD is available at https://github.com/hly1998/BRCD.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (66)
  1. Hashnet: Deep learning to hash by continuation. In Proceedings of the IEEE international conference on computer vision. 5608–5617.
  2. Junjie Chen and William K Cheung. 2019. Similarity preserving deep asymmetric quantization for image retrieval. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33. 8183–8190.
  3. A Simple Framework for Contrastive Learning of Visual Representations. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event (Proceedings of Machine Learning Research, Vol. 119). PMLR, 1597–1607.
  4. A two-step cross-modal hashing by exploiting label correlations and preserving similarity in both steps. In Proceedings of the 27th ACM International Conference on Multimedia. 1694–1702.
  5. Multi-user reinforcement learning based task migration in mobile edge computing. Frontiers of Computer Science 18, 4 (2024), 184504.
  6. Stochastic generative hashing. In International Conference on Machine Learning. PMLR, 913–922.
  7. Unsupervised deep generative adversarial hashing network. In Proceedings of the IEEE conference on computer vision and pattern recognition. 3664–3673.
  8. Khoa D Doan and Chandan K Reddy. 2020. Efficient implicit unsupervised text hashing using adversarial autoencoder. In Proceedings of The Web Conference 2020. 684–694.
  9. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In International Conference on Learning Representations.
  10. Deep Polarized Network for Supervised Learning of Accurate Binary Hashing Codes.. In IJCAI. 825–831.
  11. Vit2hash: unsupervised information-preserving hashing. arXiv preprint arXiv:2201.05541 (2022).
  12. Knowledge distillation: A survey. International Journal of Computer Vision 129 (2021), 1789–1819.
  13. Unsupervised multi-index semantic hashing. In Proceedings of the Web Conference 2021. 2879–2889.
  14. Compact hashing with joint optimization of search accuracy and time. In CVPR 2011. IEEE, 753–760.
  15. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. 770–778.
  16. An Efficient and Robust Semantic Hashing Framework for Similar Text Search. ACM Transactions on Information Systems 41, 4 (2023), 1–31.
  17. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015).
  18. Pseudo label based unsupervised deep discriminative hashing for image retrieval. In Proceedings of the 25th ACM international conference on Multimedia. 1584–1590.
  19. Young Kyun Jang and Nam Ik Cho. 2021. Self-supervised product quantization for deep unsupervised image retrieval. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 12085–12094.
  20. Deep hash distillation for image retrieval. In European Conference on Computer Vision. Springer, 354–371.
  21. Learning multiple layers of features from tiny images. (2009).
  22. Content-based multimedia information retrieval: State of the art and challenges. ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM) 2, 1 (2006), 1–19.
  23. Yunqiang Li and Jan van Gemert. 2021. Deep unsupervised image hashing by maximizing bit entropy. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35. 2002–2010.
  24. A general two-step approach to learning-based hashing. In Proceedings of the IEEE international conference on computer vision. 2552–2559.
  25. Microsoft coco: Common objects in context. In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13. Springer, 740–755.
  26. Guiding Mathematical Reasoning via Mastering Commonsense Formula Knowledge. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 1477–1488.
  27. Supervised discrete hashing with mutual linear regression. In Proceedings of the 27th ACM International Conference on Multimedia. 1561–1568.
  28. Knowledge Distillation for High Dimensional Search Index. Advances in Neural Information Processing Systems 36 (2024).
  29. A survey on deep hashing methods. ACM Transactions on Knowledge Discovery from Data 17, 1 (2023), 1–50.
  30. Deep unsupervised hashing by global and local consistency. In 2021 IEEE International Conference on Multimedia and Expo (ICME). IEEE, 1–6.
  31. Cimon: Towards high-quality hash codes. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, Virtual Event / Montreal, Canada, 19-27 August 2021. 902–908.
  32. Robust discrete code modeling for supervised hashing. Pattern Recognition 75 (2018), 128–135.
  33. Improved knowledge distillation via teacher assistant. In Proceedings of the AAAI conference on artificial intelligence, Vol. 34. 5191–5198.
  34. Fast search in hamming space with multi-index hashing. In 2012 IEEE conference on computer vision and pattern recognition. IEEE, 3108–3115.
  35. Relational knowledge distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 3967–3976.
  36. Nikolaos Passalis and Anastasios Tefas. 2018. Learning deep representations with probabilistic knowledge transfer. In Proceedings of the European Conference on Computer Vision (ECCV). 268–284.
  37. Unsupervised hashing with contrastive information bottleneck. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, Virtual Event / Montreal, Canada, 19-27 August 2021. 959–965.
  38. Peter J Rousseeuw. 1987. Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. Journal of computational and applied mathematics 20 (1987), 53–65.
  39. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition. 4510–4520.
  40. A Theoretical Analysis of Contrastive Unsupervised Representation Learning. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA (Proceedings of Machine Learning Research, Vol. 97), Kamalika Chaudhuri and Ruslan Salakhutdinov (Eds.). PMLR, 5628–5637.
  41. Unsupervised binary representation learning with deep variational networks. International Journal of Computer Vision 127, 11-12 (2019), 1614–1628.
  42. Auto-encoding twin-bottleneck hashing. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2818–2827.
  43. Karen Simonyan and Andrew Zisserman. 2015. Very deep convolutional networks for large-scale image recognition. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings.
  44. Deep robust multilevel semantic hashing for multi-label cross-modal retrieval. Pattern Recognition 120 (2021), 108084.
  45. Binary generative adversarial networks for image retrieval. In Proceedings of the AAAI conference on artificial intelligence, Vol. 32.
  46. Greedy hash: Towards fast optimization for accurate hash coding in cnn. Advances in neural information processing systems 31 (2018).
  47. HEART: Towards Effective Hash Codes under Label Noise. In Proceedings of the 30th ACM International Conference on Multimedia. 366–375.
  48. Mingxing Tan and Quoc Le. 2019. Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning. PMLR, 6105–6114.
  49. Contrastive representation distillation. arXiv preprint arXiv:1910.10699 (2019).
  50. Efficient similar exercise retrieval model based on unsupervised semantic hashing. Journal of Computer Applications 44, 1 (2024), 206.
  51. Partial-softmax loss based deep hashing. In Proceedings of the Web Conference 2021. 2869–2878.
  52. Frederick Tung and Greg Mori. 2019. Similarity-preserving knowledge distillation. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 1365–1374.
  53. Contrastive quantization with code memory for unsupervised image retrieval. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 36. 2468–2476.
  54. A survey on learning to hash. IEEE transactions on pattern analysis and machine intelligence 40, 4 (2017), 769–790.
  55. Privileged graph distillation for cold start recommendation. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval. 1187–1196.
  56. Knowledge distillation meets self-supervision. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part IX. Springer, 588–604.
  57. Semantic structure-based unsupervised deep hashing. In Proceedings of the 27th international joint conference on artificial intelligence. 1064–1070.
  58. Distillhash: Unsupervised deep hashing by distilling data pairs. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2946–2955.
  59. Pay Attention to Your Positive Pairs: Positive Pair Aware Contrastive Knowledge Distillation. In Proceedings of the 30th ACM International Conference on Multimedia. 5862–5870.
  60. Central similarity quantization for efficient image and video retrieval. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 3083–3092.
  61. Supervised hierarchical deep hashing for cross-modal retrieval. In Proceedings of the 28th ACM International Conference on Multimedia. 3386–3394.
  62. Unsupervised deep hashing with pseudo labels for scalable image retrieval. IEEE Transactions on Image Processing 27, 4 (2017), 1626–1638.
  63. Deep unsupervised hybrid-similarity hadamard hashing. In Proceedings of the 28th ACM International Conference on Multimedia. 3274–3282.
  64. Simulating Student Interactions with Two-stage Imitation Learning for Intelligent Educational Systems. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. 3423–3432.
  65. Deep balanced discrete hashing for image retrieval. Neurocomputing 403 (2020), 224–236.
  66. Complementary relation contrastive distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 9260–9269.
Citations (4)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com