Zero-shot Classification using Hyperdimensional Computing (2401.16876v1)
Abstract: Classification based on Zero-shot Learning (ZSL) is the ability of a model to classify inputs into novel classes on which the model has not previously seen any training examples. Providing an auxiliary descriptor in the form of a set of attributes describing the new classes involved in the ZSL-based classification is one of the favored approaches to solving this challenging task. In this work, inspired by Hyperdimensional Computing (HDC), we propose the use of stationary binary codebooks of symbol-like distributed representations inside an attribute encoder to compactly represent a computationally simple end-to-end trainable model, which we name Hyperdimensional Computing Zero-shot Classifier~(HDC-ZSC). It consists of a trainable image encoder, an attribute encoder based on HDC, and a similarity kernel. We show that HDC-ZSC can be used to first perform zero-shot attribute extraction tasks and, can later be repurposed for Zero-shot Classification tasks with minimal architectural changes and minimal model retraining. HDC-ZSC achieves Pareto optimal results with a 63.8% top-1 classification accuracy on the CUB-200 dataset by having only 26.6 million trainable parameters. Compared to two other state-of-the-art non-generative approaches, HDC-ZSC achieves 4.3% and 9.9% better accuracy, while they require more than 1.85x and 1.72x parameters compared to HDC-ZSC, respectively.
- O. Vinyals, C. Blundell, T. Lillicrap et al., “Matching networks for one shot learning,” Advances in Neural Information Processing Systems (NeurIPS), vol. 29, 2016.
- J. Snell, K. Swersky, and R. Zemel, “Prototypical networks for few-shot learning,” Advances in Neural Information Processing Systems (NeurIPS), vol. 30, 2017.
- V. Sunder, A. Srinivasan, L. Vig et al., “One-shot information extraction from document images using neuro-deductive program synthesis,” International Workshop on Neural-Symbolic Learning and Reasoning (NeSy), 2019.
- B. Romera-Paredes and P. Torr, “An embarrassingly simple approach to zero-shot learning,” in Proceedings of the 32nd International Conference on Machine Learning (ICML), vol. 37. PMLR, 2015, pp. 2152–2161.
- E. Altszyler, P. Brusco, N. Basiou et al., “Zero-shot multi-domain dialog state tracking using descriptive rules,” International Workshop on Neural-Symbolic Learning and Reasoning (NeSy), 2020.
- M. Norouzi, T. Mikolov, S. Bengio et al., “Zero-shot learning by convex combination of semantic embeddings,” arXiv preprint arXiv:1312.5650, 2013.
- E. Kodirov, T. Xiang, and S. Gong, “Semantic autoencoder for zero-shot learning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 3174–3183.
- Z. Zhang and V. Saligrama, “Zero-shot learning via semantic similarity embedding,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), December 2015.
- D. Huynh and E. Elhamifar, “Compositional fine-grained low-shot learning,” arXiv preprint arXiv:2105.10438, 2021.
- S. Narayan, A. Gupta, F. S. Khan et al., “Latent embedding feedback and discriminative features for zero-shot classification,” in European Conference on Computer Vision (ECCV). Springer, 2020, pp. 479–495.
- A. Radford, J. W. Kim, C. Hallacy et al., “Learning transferable visual models from natural language supervision,” in International Conference on Machine Learning (ICML). PMLR, 2021, pp. 8748–8763.
- J. Li, M. Jing, K. Lu et al., “Leveraging the invariant side of generative zero-shot learning,” in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 7394–7403.
- P. Kanerva, “Hyperdimensional computing: An introduction to computing in distributed representation with high-dimensional random vectors,” Cognitive Computation, vol. 1, no. 2, pp. 139–159, 2009.
- G. Karunaratne, M. Schmuck, M. Le Gallo et al., “Robust high-dimensional memory-augmented neural networks,” Nature Communications, vol. 12, no. 1, pp. 1–12, 2021.
- M. Hersche, G. Karunaratne, G. Cherubini et al., “Constrained Few-shot Class-incremental Learning,” in Conference on Computer Vision and Pattern Recognition (CVPR), 2022.
- G. Karunaratne, M. Hersche, J. Langeneager et al., “In-memory realization of in-situ few-shot continual learning with a dynamically evolving explicit memory,” in IEEE 48th European Solid State Circuits Conference (ESSCIRC), 2022, pp. 105–108.
- H. Jiang, R. Wang, S. Shan et al., “Transferable contrastive network for generalized zero-shot learning,” CoRR, vol. abs/1908.05832, 2019.
- P. D. Schiavone, D. Rossi, A. Pullini et al., “Quentin: an ultra-low-power pulpissimo soc in 22nm fdx,” in 2018 IEEE SOI-3D-Subthreshold Microelectronics Technology Unified Conference (S3S), 2018, pp. 1–3.
- Y. Xian, C. H. Lampert, B. Schiele et al., “Zero-shot learning—a comprehensive evaluation of the good, the bad and the ugly,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 41, no. 9, pp. 2251–2265, 2018.
- Z. Akata, F. Perronnin, Z. Harchaoui et al., “Label-embedding for image classification,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 7, pp. 1425–1438, 2015.
- Y. Xian, Z. Akata, G. Sharma et al., “Latent embeddings for zero-shot classification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 69–77.
- R. Socher, M. Ganjoo, C. D. Manning et al., “Zero-shot learning through cross-modal transfer,” Advances in Neural Information Processing Systems (NeurIPS), vol. 26, 2013.
- C. H. Lampert, H. Nickisch, and S. Harmeling, “Attribute-based classification for zero-shot visual object categorization,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 36, pp. 453–465, 2014.
- S. Changpinyo, W.-L. Chao, B. Gong et al., “Synthesized classifiers for zero-shot learning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 5327–5336.
- V. K. Verma and P. Rai, “A simple exponential family framework for zero-shot learning,” in Machine Learning and Knowledge Discovery in Databases. Springer International Publishing, 2017, pp. 792–808.
- Y. Xian, S. Sharma, B. Schiele et al., “F-VAEGAN-D2: A feature generating framework for any-shot learning,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- R. Felix, I. Reid, G. Carneiro et al., “Multi-modal cycle-consistent generalized zero-shot learning,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 21–37.
- Y. Xian, T. Lorenz, B. Schiele et al., “Feature generating networks for zero-shot learning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 5542–5551.
- A. Rahimi, P. Kanerva, and J. M. Rabaey, “A robust and energy-efficient classifier using brain-inspired hyperdimensional computing,” in Proceedings of the 2016 International Symposium on Low Power Electronics and Design (ISLPED), 2016.
- M. Schmuck, L. Benini, and A. Rahimi, “Hardware optimizations of dense binary hyperdimensional computing: Rematerialization of hypervectors, binarized bundling, and combinational associative memory,” ACM Journal on Emerging Technologies in Computing Systems (JETC), vol. 15, no. 4, pp. 1–25, 2019.
- C. Wah, S. Branson, P. Welinder et al., “Cub-200 2011 dataset,” California Institute of Technology, Tech. Rep. CNS-TR-2011-001, 2011.
- I. Loshchilov and F. Hutter, “Decoupled weight decay regularization,” in International Conference on Learning Representations (ICLR), 2019.
- ——, “SGDR: Stochastic gradient descent with warm restarts,” arXiv preprint arXiv:1608.03983, 2016.
- R. Zakizadeh, M. Sasdelli, Y. Qian et al., “Finetag: Multi-attribute classification at fine-grained level in images,” arXiv preprint arXiv:1806.07124, 2018.
- K. Han, J. Guo, C. Zhang et al., “Attribute-aware attention model for fine-grained representation learning,” in Proceedings of the 26th ACM international conference on Multimedia, 2018, pp. 2040–2048.
- A. Garofalo, G. Ottavi, F. Conti et al., “A heterogeneous in-memory computing cluster for flexible end-to-end inference of real-world deep neural networks,” IEEE Journal on Emerging and Selected Topics in Circuits and Systems (JETCAS), vol. 12, no. 2, pp. 422–435, 2022.
- R. Khaddam-Aljameh, M. Stanisavljevic, J. F. Mas et al., “Hermes-core—a 1.59-tops/mm 2 pcm on 14-nm cmos in-memory compute core using 300-ps/lsb linearized cco-based adcs,” IEEE Journal of Solid-State Circuits, vol. 57, no. 4, pp. 1027–1038, 2022.
- M. Eggimann, A. Rahimi, and L. Benini, “A 5 μ𝜇\muitalic_μw standard cell memory-based configurable hyperdimensional computing accelerator for always-on smart sensing,” IEEE Transactions on Circuits and Systems I: Regular Papers, vol. 68, no. 10, pp. 4116–4128, 2021.