Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
144 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Stability Plasticity Decoupled Fine-tuning For Few-shot end-to-end Object Detection (2401.11140v1)

Published 20 Jan 2024 in cs.CV and cs.AI

Abstract: Few-shot object detection(FSOD) aims to design methods to adapt object detectors efficiently with only few annotated samples. Fine-tuning has been shown to be an effective and practical approach. However, previous works often take the classical base-novel two stage fine-tuning procedure but ignore the implicit stability-plasticity contradiction among different modules. Specifically, the random re-initialized classifiers need more plasticity to adapt to novel samples. The other modules inheriting pre-trained weights demand more stability to reserve their class-agnostic knowledge. Regular fine-tuning which couples the optimization of these two parts hurts the model generalization in FSOD scenarios. In this paper, we find that this problem is prominent in the end-to-end object detector Sparse R-CNN for its multi-classifier cascaded architecture. We propose to mitigate this contradiction by a new three-stage fine-tuning procedure by introducing an addtional plasticity classifier fine-tuning(PCF) stage. We further design the multi-source ensemble(ME) technique to enhance the generalization of the model in the final fine-tuning stage. Extensive experiments verify that our method is effective in regularizing Sparse R-CNN, outperforming previous methods in the FSOD benchmark.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. “Sparse r-cnn: End-to-end object detection with learnable proposals,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 14454–14463.
  2. “Faster r-cnn: Towards real-time object detection with region proposal networks,” Advances in neural information processing systems, vol. 28, 2015.
  3. “Frustratingly simple few-shot object detection,” 37th International Conference on Machine Learning, ICML 2020, vol. PartF168147-13, pp. 9861–9870, 2020.
  4. “Multi-scale positive sample refinement for few-shot object detection,” in European conference on computer vision. Springer, 2020, pp. 456–472.
  5. “FSCE: Few-Shot Object Detection via Contrastive Proposal Encoding,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 7348–7358, 2021.
  6. “Few-shot object detection via association and discrimination,” Advances in Neural Information Processing Systems, vol. 34, 2021.
  7. “End-to-end object detection with transformers,” in European conference on computer vision. Springer, 2020, pp. 213–229.
  8. “A comprehensive survey of continual learning: Theory, method and application,” arXiv preprint arXiv:2302.00487, 2023.
  9. “Fine-tuning can distort pretrained features and underperform out-of-distribution,” in International Conference on Learning Representations, 2021.
  10. “Few-shot object detection via feature reweighting,” Proceedings of the IEEE International Conference on Computer Vision, vol. 2019-October, pp. 8419–8428, 2019.
  11. “Meta R-CNN: Towards general solver for instance-level low-shot learning,” Proceedings of the IEEE International Conference on Computer Vision, vol. 2019-October, pp. 9576–9585, 2019.
  12. “Few-shot object detection and viewpoint estimation for objects in the wild,” in European conference on computer vision. Springer, 2020, pp. 192–210.
  13. “Meta-detr: Few-shot object detection via unified image-level meta-learning,” arXiv preprint arXiv:2103.11731, vol. 2, no. 6, 2021.
  14. “DeFRCN: Decoupled Faster R-CNN for Few-Shot Object Detection,” ICCV2021, 2021.
  15. “Deformable detr: Deformable transformers for end-to-end object detection,” in International Conference on Learning Representations, 2020.
  16. “Few-shot end-to-end object detection via constantly concentrated encoding across heads,” in European Conference on Computer Vision. Springer, 2022, pp. 57–73.
  17. “Continual lifelong learning with neural networks: A review,” Neural networks, vol. 113, pp. 54–71, 2019.
  18. “Overcoming catastrophic forgetting in neural networks,” Proceedings of the national academy of sciences, vol. 114, no. 13, pp. 3521–3526, 2017.
  19. “Memory aware synapses: Learning what (not) to forget,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 139–154.
  20. “Nispa: Neuro-inspired stability-plasticity adaptation for continual learning in sparse networks,” arXiv preprint arXiv:2206.09117, 2022.
  21. “Coscl: Cooperation of small continual learners is stronger than a big one,” in European Conference on Computer Vision. Springer, 2022, pp. 254–271.
  22. “Overcoming catastrophic forgetting with hard attention to the task,” in International conference on machine learning. PMLR, 2018, pp. 4548–4557.
  23. “Algorithmic regularization in learning deep homogeneous models: Layers are automatically balanced,” Advances in neural information processing systems, vol. 31, 2018.
  24. “Prototypical networks for few-shot learning,” Advances in Neural Information Processing Systems, vol. 2017-December, pp. 4078–4088, 2017.
  25. “Robust fine-tuning of zero-shot models,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 7959–7971.
  26. “Linear mode connectivity and the lottery ticket hypothesis,” in International Conference on Machine Learning. PMLR, 2020, pp. 3259–3269.
  27. “Few-shot object detection with attention-rpn and multi-relation detector,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 4013–4022.
  28. “Transformation invariant few-shot object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 3094–3102.
  29. “Meta faster r-cnn: Towards accurate few-shot object detection with attentive feature alignment,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2022, vol. 36, pp. 780–789.
  30. “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets