Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Boosting Few-Shot Learning with Disentangled Self-Supervised Learning and Meta-Learning for Medical Image Classification (2403.17530v1)

Published 26 Mar 2024 in cs.CV and cs.AI

Abstract: Background and objective: Employing deep learning models in critical domains such as medical imaging poses challenges associated with the limited availability of training data. We present a strategy for improving the performance and generalization capabilities of models trained in low-data regimes. Methods: The proposed method starts with a pre-training phase, where features learned in a self-supervised learning setting are disentangled to improve the robustness of the representations for downstream tasks. We then introduce a meta-fine-tuning step, leveraging related classes between meta-training and meta-testing phases but varying the granularity level. This approach aims to enhance the model's generalization capabilities by exposing it to more challenging classification tasks during meta-training and evaluating it on easier tasks but holding greater clinical relevance during meta-testing. We demonstrate the effectiveness of the proposed approach through a series of experiments exploring several backbones, as well as diverse pre-training and fine-tuning schemes, on two distinct medical tasks, i.e., classification of prostate cancer aggressiveness from MRI data and classification of breast cancer malignity from microscopic images. Results: Our results indicate that the proposed approach consistently yields superior performance w.r.t. ablation experiments, maintaining competitiveness even when a distribution shift between training and evaluation data occurs. Conclusion: Extensive experiments demonstrate the effectiveness and wide applicability of the proposed approach. We hope that this work will add another solution to the arsenal of addressing learning issues in data-scarce imaging domains.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (59)
  1. Generalizing from a few examples: A survey on few-shot learning. ACM computing surveys (csur), 53(3):1–34, 2020. doi:https://doi.org/10.1145/3386252.
  2. Causality-driven one-shot learning for prostate cancer grading from mri. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2616–2624, 2023. doi:10.1109/ICCVW60793.2023.00276.
  3. Pfemed: Few-shot medical image classification using prior guided feature enhancement. Pattern Recognition, 134:109108, 2023. doi:https://doi.org/10.1016/j.patcog.2022.109108.
  4. Multi-Learner Based Deep Meta-Learning for Few-Shot Medical Image Classification. IEEE Journal of Biomedical and Health Informatics, 27(1):17–28, 2022. doi:10.1109/JBHI.2022.3215147. Publisher: IEEE.
  5. A systematic review of few-shot learning in medical imaging. arXiv, 2023. doi:https://doi.org/10.48550/arXiv.2309.11433.
  6. A closer look at few-shot classification. arXiv, 2019. doi:https://doi.org/10.48550/arXiv.1904.04232.
  7. Boosting few-shot visual learning with self-supervision. In Proceedings of the IEEE/CVF international conference on computer vision, pages 8059–8068, 2019. doi:10.1109/ICCV.2019.00815.
  8. When does self-supervision improve few-shot learning? In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part VII 16, pages 645–666. Springer, 2020. doi:https://doi.org/10.1007/978-3-030-58571-6_38.
  9. Self-supervised learning disentangled group representation as feature. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 18225–18240. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper_files/paper/2021/file/97416ac0f58056947e2eb5d5d253d4f2-Paper.pdf.
  10. Joint distribution matters: Deep brownian distance covariance for few-shot classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7972–7981, 2022. doi:10.1109/CVPR52688.2022.00781.
  11. Artificial intelligence and radiologists at prostate cancer detection in mri—the pi-cai challenge. In Medical Imaging with Deep Learning, short paper track, 2023.
  12. A dataset for breast cancer histopathological image classification. Ieee transactions on biomedical engineering, 63(7):1455–1462, 2015. doi:10.1109/TBME.2015.2496264.
  13. International Society of Urological Pathology (ISUP) grading of prostate cancer - An ISUP consensus on contemporary grading. APMIS: acta pathologica, microbiologica, et immunologica Scandinavica, 124(6):433–435, June 2016. ISSN 1600-0463. doi:10.1111/apm.12533.
  14. Early detection of prostate cancer in 2020 and beyond: Facts and recommendations for the european union and the european commission. European Urology, 79(3):327–329, 2021. ISSN 0302-2838. doi:https://doi.org/10.1016/j.eururo.2020.12.010. URL https://www.sciencedirect.com/science/article/pii/S0302283820309581.
  15. Computer assisted recognition of breast cancer in biopsy images via fusion of nucleus-guided deep convolutional features. Computer Methods and Programs in Biomedicine, 194:105531, 2020. ISSN 0169-2607. doi:https://doi.org/10.1016/j.cmpb.2020.105531. URL https://www.sciencedirect.com/science/article/pii/S0169260720302911.
  16. Breakhis based breast cancer automatic diagnosis using deep learning: Taxonomy, survey and insights. Neurocomputing, 375:9–24, 2020. ISSN 0925-2312. doi:https://doi.org/10.1016/j.neucom.2019.09.044. URL https://www.sciencedirect.com/science/article/pii/S0925231219313128.
  17. MetaMed: Few-shot medical image classification using gradient-based meta-learning. Pattern Recognition, 120:108111, 2021. doi:https://doi.org/10.1016/j.patcog.2021.108111. Publisher: Elsevier.
  18. Self-supervised learning for few-shot image classification. In ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1745–1749, 2021. doi:10.1109/ICASSP39728.2021.9413783.
  19. Self-supervised vision transformer-based few-shot learning for facial expression recognition. Information Sciences, 634:206–226, 2023. doi:https://doi.org/10.1016/j.ins.2023.03.105.
  20. Self-supervised prototypical transfer learning for few-shot classification. arXiv, 2020. doi:https://doi.org/10.48550/arXiv.2006.11325.
  21. Few-shot classification with contrastive learning. In European Conference on Computer Vision, pages 293–309. Springer, 2022. doi:https://doi.org/10.1007/978-3-031-20044-1_17.
  22. Task-adaptive feature disentanglement and hallucination for few-shot classification. IEEE Transactions on Circuits and Systems for Video Technology, 2023. doi:10.1109/TCSVT.2023.3238804.
  23. Joint feature disentanglement and hallucination for few-shot image classification. IEEE Transactions on Image Processing, 30:9245–9258, 2021. doi:10.1109/TIP.2021.3124322.
  24. Disentangled feature representation for few-shot image classification. IEEE Transactions on Neural Networks and Learning Systems, 2023. doi:10.1109/TNNLS.2023.3241919.
  25. Multi-view perceptron: a deep model for learning face identity and view representations. In Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 27. Curran Associates, Inc., 2014. URL https://proceedings.neurips.cc/paper_files/paper/2014/file/140f6969d5213fd0ece03148e62e461e-Paper.pdf.
  26. Learning to decompose and disentangle representations for video prediction. CoRR, abs/1806.04166, 2018. doi:https://doi.org/10.48550/arXiv.1806.04166.
  27. Learning disentangled semantic representation for domain adaptation. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, pages 2060–2066. International Joint Conferences on Artificial Intelligence Organization, 7 2019. doi:10.24963/ijcai.2019/285. URL https://doi.org/10.24963/ijcai.2019/285.
  28. Learning to disentangle factors of variation with manifold interaction. In Eric P. Xing and Tony Jebara, editors, Proceedings of the 31st International Conference on Machine Learning, volume 32 of Proceedings of Machine Learning Research, pages 1431–1439, Bejing, China, 22–24 Jun 2014. PMLR. URL https://proceedings.mlr.press/v32/reed14.html.
  29. Bayesian representation learning with oracle constraints. arXiv, 2015. doi:https://doi.org/10.48550/arXiv.1506.05011.
  30. Learning disentangled representations in the imaging domain. Medical Image Analysis, 80:102516, 2022. doi:https://doi.org/10.1016/j.media.2022.102516.
  31. Challenging common assumptions in the unsupervised learning of disentangled representations. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 4114–4124. PMLR, 09–15 Jun 2019. URL https://proceedings.mlr.press/v97/locatello19a.html.
  32. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. URL https://proceedings.neurips.cc/paper_files/paper/2016/file/7c9d0b1f96aebd7b5eca8c3edaa19ebb-Paper.pdf.
  33. InfoGAN-CR and ModelCentrality: Self-supervised model training and selection for disentangling GANs. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 6127–6139. PMLR, 13–18 Jul 2020. URL https://proceedings.mlr.press/v119/lin20e.html.
  34. Elastic-infogan: Unsupervised disentangled representation learning in class-imbalanced data. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 18063–18075. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/d1e39c9bda5c80ac3d8ea9d658163967-Paper.pdf.
  35. Isolating sources of disentanglement in variational autoencoders. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. URL https://proceedings.neurips.cc/paper_files/paper/2018/file/1ee3dfcd8a0645a25a35977997223d22-Paper.pdf.
  36. Disentangling by factorising. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 2649–2658. PMLR, 10–15 Jul 2018. URL https://proceedings.mlr.press/v80/kim18b.html.
  37. beta-vae: Learning basic visual concepts with a constrained variational framework. In International conference on learning representations, 2016.
  38. Invariant risk minimization. arXiv, 2019. doi:10.48550/arXiv.1907.02893.
  39. Dive into the details of self-supervised learning for medical image analysis. Medical Image Analysis, 89:102879, 2023. doi:https://doi.org/10.1016/j.media.2023.102879.
  40. Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems, 33:21271–21284, 2020.
  41. Unsupervised visual representation learning by context prediction. In Proceedings of the IEEE international conference on computer vision, pages 1422–1430, 2015. doi:https://doi.org/10.1109/ICCV.2015.167.
  42. A simple framework for contrastive learning of visual representations. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 1597–1607. PMLR, 13–18 Jul 2020. URL https://proceedings.mlr.press/v119/chen20j.html.
  43. James Woodward. Making things happen: A theory of causal explanation. Oxford university press, 2005.
  44. V. Vapnik. Principles of risk minimization for learning theory. In J. Moody, S. Hanson, and R.P. Lippmann, editors, Advances in Neural Information Processing Systems, volume 4. Morgan-Kaufmann, 1991. URL https://proceedings.neurips.cc/paper_files/paper/1991/file/ff4d5fbbafdf976cfdc032e3bde78de5-Paper.pdf.
  45. Prototypical networks for few-shot learning. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL https://proceedings.neurips.cc/paper_files/paper/2017/file/cb8da6767461f2812ae4290eac7cbc42-Paper.pdf.
  46. Few-shot learning with localization in realistic settings. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 6558–6567, 2019. doi:10.1109/CVPR.2019.00672.
  47. Asymmetric distribution measure for few-shot learning. arXiv, 2020. doi:https://doi.org/10.48550/arXiv.2002.00153.
  48. Pytorch: An imperative style, high-performance deep learning library. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL https://proceedings.neurips.cc/paper_files/paper/2019/file/bdbca288fee7f92f2bfa9f7012727740-Paper.pdf.
  49. Meta-dataset: A dataset of datasets for learning to learn from few examples. arXiv, 2019. doi:https://doi.org/10.48550/arXiv.1903.03096.
  50. Matching networks for one shot learning. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. URL https://proceedings.neurips.cc/paper_files/paper/2016/file/90e1357833654983612fb05e3ec9148c-Paper.pdf.
  51. Mayke Pereira. Breakhis - breast cancer histopathological database, 2023.
  52. Two sides of meta-learning evaluation: In vs. out of distribution. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 3770–3783. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper_files/paper/2021/file/1e932f24dc0aa4e7a6ac2beec387416d-Paper.pdf.
  53. Large-scale robust deep auc maximization: A new surrogate loss and empirical studies on medical image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3040–3049, 2021. doi:10.1109/ICCV48922.2021.00303.
  54. Libauc: A deep learning library for x-risk optimization. In 29th SIGKDD Conference on Knowledge Discovery and Data Mining, 2023.
  55. Fast objective & duality gap convergence for non-convex strongly-concave min-max problems with pl condition. Journal of Machine Learning Research, 24:1–63, 2023.
  56. Wheater’s functional histology E-Book: a text and colour atlas. Elsevier Health Sciences, 2013.
  57. Histology image analysis for carcinoma detection and grading. Computer methods and programs in biomedicine, 107(3):538–556, 2012. doi:https://doi.org/10.1016/j.cmpb.2011.12.007.
  58. Bridging the gap between prostate radiology and pathology through machine learning. Medical Physics, 49(8):5160–5181, 2022. doi:https://doi.org/10.1002/mp.15777.
  59. Prostate imaging features that indicate benign or malignant pathology on biopsy. Translational andrology and urology, 7(Suppl 4):S420, 2018. doi:10.21037/tau.2018.07.06.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com