Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 69 tok/s
Gemini 2.5 Pro 52 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 28 tok/s Pro
GPT-4o 75 tok/s Pro
Kimi K2 191 tok/s Pro
GPT OSS 120B 402 tok/s Pro
Claude Sonnet 4.5 34 tok/s Pro
2000 character limit reached

A differentiable Gaussian Prototype Layer for explainable Segmentation (2306.14361v1)

Published 25 Jun 2023 in cs.CV

Abstract: We introduce a Gaussian Prototype Layer for gradient-based prototype learning and demonstrate two novel network architectures for explainable segmentation one of which relies on region proposals. Both models are evaluated on agricultural datasets. While Gaussian Mixture Models (GMMs) have been used to model latent distributions of neural networks before, they are typically fitted using the EM algorithm. Instead, the proposed prototype layer relies on gradient-based optimization and hence allows for end-to-end training. This facilitates development and allows to use the full potential of a trainable deep feature extractor. We show that it can be used as a novel building block for explainable neural networks. We employ our Gaussian Prototype Layer in (1) a model where prototypes are detected in the latent grid and (2) a model inspired by Fast-RCNN with SLIC superpixels as region proposals. The earlier achieves a similar performance as compared to the state-of-the art while the latter has the benefit of a more precise prototype localization that comes at the cost of slightly lower accuracies. By introducing a gradient-based GMM layer we combine the benefits of end-to-end training with the simplicity and theoretical foundation of GMMs which will allow to adapt existing semi-supervised learning strategies for prototypical part models in future.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. “This looks like that: deep learning for interpretable image recognition,” Advances in neural information processing systems, vol. 32, 2019.
  2. “But that’s not why: Inference adjustment by interactive prototype deselection,” arXiv preprint arXiv:2203.10087, 2022.
  3. “Towards scalable and unified example-based explanation and outlier detection,” IEEE Transactions on Image Processing, 2021.
  4. “Interpretable image classification with differentiable prototypes assignment,” in Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XII. Springer, 2022, pp. 351–368.
  5. “Semi-protopnet deep neural network for the classification of defective power grid distribution structures,” Sensors, vol. 22, no. 13, pp. 4859, 2022.
  6. “These do not look like those: An interpretable deep learning model for image recognition,” IEEE Access, vol. 9, pp. 41482–41493, 2021.
  7. “Protgnn: Towards self-explaining graph neural networks,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2022, vol. 36, pp. 9127–9135.
  8. “Vit-net: Interpretable vision transformers with neural tree decoder,” in International Conference on Machine Learning. PMLR, 2022, pp. 11162–11172.
  9. “Interpretable image recognition with hierarchical prototypes,” in Proceedings of the AAAI Conference on Human Computation and Crowdsourcing, 2019, vol. 7, pp. 32–40.
  10. “Deformable protopnet: An interpretable image classifier using deformable prototypes,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 10265–10275.
  11. “Few-shot semantic segmentation with prototype learning.,” in BMVC, 2018, vol. 3.
  12. “Prototypical cross-attention networks for multiple object tracking and segmentation,” Advances in Neural Information Processing Systems, vol. 34, pp. 1192–1203, 2021.
  13. “Gradient-based training of gaussian mixture models for high-dimensional streaming data,” Neural Processing Letters, vol. 53, no. 6, pp. 4331–4348, 2021.
  14. “This looks more like that: Enhancing self-explaining models by prototypical relevance propagation,” Pattern Recognition, vol. 136, pp. 109172, 2023.
  15. “Semi-supervised classification based on gaussian mixture model for remote imagery,” Science China Technological Sciences, vol. 53, no. 1, pp. 85–90, 2010.
  16. Ross Girshick, “Fast r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 1440–1448.
  17. “Slic superpixels,” Tech. Rep., 2010.
  18. “Mask r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2961–2969.
  19. “Minneapple: A benchmark dataset for apple detection and segmentation,” 2019.
  20. “Grape detection, segmentation, and tracking using deep neural networks and three-dimensional association,” Computers and Electronics in Agriculture, vol. 170, pp. 105247, 2020.
  21. “Minneapple: a benchmark dataset for apple detection and segmentation,” IEEE Robotics and Automation Letters, vol. 5, no. 2, pp. 852–858, 2020.
Citations (4)

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube