Structured Model Pruning for Efficient Inference in Computational Pathology (2404.08831v1)
Abstract: Recent years have seen significant efforts to adopt AI in healthcare for various use cases, from computer-aided diagnosis to ICU triage. However, the size of AI models has been rapidly growing due to scaling laws and the success of foundational models, which poses an increasing challenge to leverage advanced models in practical applications. It is thus imperative to develop efficient models, especially for deploying AI solutions under resource-constrains or with time sensitivity. One potential solution is to perform model compression, a set of techniques that remove less important model components or reduce parameter precision, to reduce model computation demand. In this work, we demonstrate that model pruning, as a model compression technique, can effectively reduce inference cost for computational and digital pathology based analysis with a negligible loss of analysis performance. To this end, we develop a methodology for pruning the widely used U-Net-style architectures in biomedical imaging, with which we evaluate multiple pruning heuristics on nuclei instance segmentation and classification, and empirically demonstrate that pruning can compress models by at least 70% with a negligible drop in performance.
- Combining weight pruning and knowledge distillation for cnn compression. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pages 3185–3192, 2021. 10.1109/CVPRW53098.2021.00356.
- Recurrent residual u-net for medical image segmentation. Journal of medical imaging, 6(1):014006–014006, 2019.
- Digital pathology operations at a tertiary cancer center: Infrastructure requirements and operational cost. Journal of Pathology Informatics, 14:100318, 2023.
- A review of neural architecture search. Neurocomputing, 474:82–93, 2022.
- What is the state of neural network pruning?, 2020.
- A survey on deep neural network Pruning-Taxonomy, comparison, analysis, and recommendations. August 2023.
- On the efficacy of knowledge distillation. In Proceedings of the IEEE/CVF international conference on computer vision, pages 4794–4802, 2019.
- A transfer learning with structured filter pruning approach for improved breast cancer classification on point-of-care devices. Computers in Biology and Medicine, 134:104432, 2021. ISSN 0010-4825. https://doi.org/10.1016/j.compbiomed.2021.104432. URL https://www.sciencedirect.com/science/article/pii/S0010482521002262.
- Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
- Segment anything model (sam) for digital pathology: Assess zero-shot segmentation on whole slide imaging. arXiv preprint arXiv:2304.04155, 2023.
- RepVGG: Making VGG-style ConvNets great again. January 2021.
- Resolution-based distillation for efficient histology image classification, 2021.
- The lottery ticket hypothesis: Finding sparse, trainable neural networks. March 2018.
- Stabilizing the lottery ticket hypothesis. March 2019.
- DepthShrinker: A new compression paradigm towards boosting Real-Hardware efficiency of compact neural networks. June 2022a.
- Depthshrinker: a new compression paradigm towards boosting real-hardware efficiency of compact neural networks. In International Conference on Machine Learning, pages 6849–6862. PMLR, 2022b.
- PanNuke dataset extension, insights and baselines. March 2020.
- A survey of quantization methods for efficient neural network inference. March 2021.
- Knowledge distillation: A survey. International Journal of Computer Vision, 129(6):1789–1819, March 2021. ISSN 1573-1405. 10.1007/s11263-021-01453-z. URL http://dx.doi.org/10.1007/s11263-021-01453-z.
- HoVer-Net: Simultaneous segmentation and classification of nuclei in Multi-Tissue histology images. December 2018.
- Deep residual learning for image recognition. December 2015.
- Identity mappings in deep residual networks. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part IV 14, pages 630–645. Springer, 2016.
- Distilling the knowledge in a neural network. March 2015.
- How does pruning impact long-tailed multi-label medical image classifiers? In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 663–673. Springer, 2023.
- Characterising bias in compressed models, 2020.
- Cellvit: Vision transformers for precise cell segmentation and classification, 2023.
- Batch normalization: Accelerating deep network training by reducing internal covariate shift. February 2015.
- Deep learning for digital pathology image analysis: A comprehensive tutorial with selected use cases. Journal of pathology informatics, 7(1):29, 2016.
- Knowledge distillation in histology landscape by multi-layer features supervision. IEEE Journal of Biomedical and Health Informatics, 27(4):2037–2046, 2023.
- Pruning adversarially robust neural networks without adversarial examples, 2022.
- Predicting survival from colorectal cancer histology slides using deep learning: A retrospective multicenter study. PLoS medicine, 16(1):e1002730, 2019.
- Adam: A method for stochastic optimization. December 2014.
- Pruning filters for efficient ConvNets. August 2016.
- Learning efficient convolutional networks through network slimming. August 2017.
- Negative instance guided self-distillation framework for whole slide image analysis. IEEE Journal of Biomedical and Health Informatics, 28(2):964–975, 2024. 10.1109/JBHI.2023.3298798.
- Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European conference on computer vision (ECCV), pages 116–131, 2018.
- Deep neural network pruning for nuclei instance segmentation in hematoxylin and eosin-stained histological images. In Applications of Medical Artificial Intelligence: First International Workshop, AMAI 2022, Held in Conjunction with MICCAI 2022, Singapore, September 18, 2022, Proceedings, page 108–117, Berlin, Heidelberg, 2022. Springer-Verlag. ISBN 978-3-031-17720-0. 10.1007/978-3-031-17721-7_12. URL https://doi.org/10.1007/978-3-031-17721-7_12.
- A comprehensive survey of neural architecture search: Challenges and solutions, 2021.
- U-net: Convolutional networks for biomedical image segmentation, 2015.
- Exploring image augmentations for siamese representation learning with chest x-rays. In Medical Imaging with Deep Learning, pages 444–467. PMLR, 2024.
- Slpd: Slide-level prototypical distillation for wsis, 2023.
- Shifting machine learning for healthcare from development to deployment and from models to data. Nature Biomedical Engineering, 6(12):1330–1345, 2022.