Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A dual-branch model with inter- and intra-branch contrastive loss for long-tailed recognition (2309.16135v1)

Published 28 Sep 2023 in cs.CV

Abstract: Real-world data often exhibits a long-tailed distribution, in which head classes occupy most of the data, while tail classes only have very few samples. Models trained on long-tailed datasets have poor adaptability to tail classes and the decision boundaries are ambiguous. Therefore, in this paper, we propose a simple yet effective model, named Dual-Branch Long-Tailed Recognition (DB-LTR), which includes an imbalanced learning branch and a Contrastive Learning Branch (CoLB). The imbalanced learning branch, which consists of a shared backbone and a linear classifier, leverages common imbalanced learning approaches to tackle the data imbalance issue. In CoLB, we learn a prototype for each tail class, and calculate an inter-branch contrastive loss, an intra-branch contrastive loss and a metric loss. CoLB can improve the capability of the model in adapting to tail classes and assist the imbalanced learning branch to learn a well-represented feature space and discriminative decision boundary. Extensive experiments on three long-tailed benchmark datasets, i.e., CIFAR100-LT, ImageNet-LT and Places-LT, show that our DB-LTR is competitive and superior to the comparative methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Long-tailed recognition via weight balancing, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6897–6907.
  2. A systematic study of the class imbalance problem in convolutional neural networks. Neural networks 106, 249–259.
  3. Ace: Ally complementary experts for solving long-tailed recognition in one-shot, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 112–121.
  4. Learning imbalanced datasets with label-distribution-aware margin loss. Advances in neural information processing systems 32.
  5. Smote: synthetic minority over-sampling technique. Journal of artificial intelligence research 16, 321–357.
  6. A simple framework for contrastive learning of visual representations, in: International conference on machine learning, PMLR. pp. 1597–1607.
  7. Remix: rebalanced mixup, in: European Conference on Computer Vision, Springer. pp. 95–110.
  8. Reslt: Residual learning for long-tailed recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence .
  9. Parametric contrastive learning, in: Proceedings of the IEEE/CVF international conference on computer vision, pp. 715–724.
  10. Class-balanced loss based on effective number of samples, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9268–9277.
  11. Imagenet: A large-scale hierarchical image database, in: 2009 IEEE conference on computer vision and pattern recognition, Ieee. pp. 248–255.
  12. C4. 5, class imbalance, and cost sensitivity: why under-sampling beats over-sampling, in: Workshop on learning from imbalanced datasets II, pp. 1–8.
  13. Momentum contrast for unsupervised visual representation learning, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9729–9738.
  14. Deep residual learning for image recognition, in: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778.
  15. Disentangling label distribution for long-tailed visual recognition, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 6626–6636.
  16. Class-distribution-aware calibration for long-tailed visual recognition. arXiv preprint arXiv:2109.05263 .
  17. Exploring balanced feature spaces for representation learning, in: International Conference on Learning Representations.
  18. Decoupling representation and classifier for long-tailed recognition, in: 8th International Conference on Learning Representations, OpenReview.net.
  19. Metasaug: Meta semantic augmentation for long-tailed visual recognition, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 5212–5221.
  20. Targeted supervised contrastive learning for long-tailed recognition, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6918–6928.
  21. Paying deep attention to both neighbors and multiple tasks, in: International Conference on Intelligent Computing, Springer. pp. 140–149.
  22. Focal loss for dense object detection, in: Proceedings of the IEEE international conference on computer vision, pp. 2980–2988.
  23. Large-scale long-tailed recognition in an open world, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2537–2546.
  24. Sgdr: Stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983 .
  25. Visualizing data using t-sne. Journal of machine learning research 9.
  26. Long-tail learning via logit adjustment, in: 9th International Conference on Learning Representations, 2021.
  27. Survey of resampling techniques for improving classification performance in unbalanced datasets. arXiv preprint arXiv:1608.06048 .
  28. Influence-balanced loss for imbalanced visual classification, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 735–744.
  29. Dynamic sampling in convolutional neural networks for imbalanced data classification, in: 2018 IEEE conference on multimedia information processing and retrieval (MIPR), IEEE. pp. 112–117.
  30. Meta-weight-net: Learning an explicit mapping for sample weighting. Advances in neural information processing systems 32.
  31. Cyclical focal loss. arXiv preprint arXiv:2202.08978 .
  32. Prototypical networks for few-shot learning. Advances in neural information processing systems 30.
  33. Rsg: A simple but effective module for learning imbalanced datasets, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3784–3793.
  34. Contrastive learning based hybrid networks for long-tailed image classification, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 943–952.
  35. Long-tailed recognition by routing diverse distribution-aware experts. arXiv preprint arXiv:2010.01809 .
  36. Implicit semantic data augmentation for deep networks. Advances in Neural Information Processing Systems 32.
  37. Dynamic auxiliary soft labels for decoupled learning. Neural Networks 151, 132–142.
  38. Learning to model the tail. Advances in neural information processing systems 30.
  39. Learning from multiple experts: Self-paced knowledge distillation for long-tailed classification, in: European Conference on Computer Vision, Springer. pp. 247–263.
  40. A survey on long-tailed visual recognition. International Journal of Computer Vision , 1–36.
  41. Cross-domain attribute representation based on convolutional neural network, in: International Conference on Intelligent Computing, Springer. pp. 134–142.
  42. mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412 .
  43. Distribution alignment: A unified framework for long-tail visual recognition, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 2361–2370.
  44. Bag of tricks for long-tailed visual recognition with deep convolutional neural networks, in: Proceedings of the AAAI conference on artificial intelligence, pp. 3447–3455.
  45. Improving calibration for long-tailed recognition, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 16489–16498.
  46. Bbn: Bilateral-branch network with cumulative learning for long-tailed visual recognition, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9719–9728.
  47. Balanced contrastive learning for long-tailed visual recognition, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6908–6917.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Qiong Chen (21 papers)
  2. Tianlin Huang (2 papers)
  3. Geren Zhu (1 paper)
  4. Enlu Lin (2 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.