Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

S-E Pipeline: A Vision Transformer (ViT) based Resilient Classification Pipeline for Medical Imaging Against Adversarial Attacks (2407.17587v1)

Published 23 Jul 2024 in cs.CV and cs.LG

Abstract: Vision Transformer (ViT) is becoming widely popular in automating accurate disease diagnosis in medical imaging owing to its robust self-attention mechanism. However, ViTs remain vulnerable to adversarial attacks that may thwart the diagnosis process by leading it to intentional misclassification of critical disease. In this paper, we propose a novel image classification pipeline, namely, S-E Pipeline, that performs multiple pre-processing steps that allow ViT to be trained on critical features so as to reduce the impact of input perturbations by adversaries. Our method uses a combination of segmentation and image enhancement techniques such as Contrast Limited Adaptive Histogram Equalization (CLAHE), Unsharp Masking (UM), and High-Frequency Emphasis filtering (HFE) as preprocessing steps to identify critical features that remain intact even after adversarial perturbations. The experimental study demonstrates that our novel pipeline helps in reducing the effect of adversarial attacks by 72.22% for the ViT-b32 model and 86.58% for the ViT-l32 model. Furthermore, we have shown an end-to-end deployment of our proposed method on the NVIDIA Jetson Orin Nano board to demonstrate its practical use case in modern hand-held devices that are usually resource-constrained.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. Q. Li, W. Cai, X. Wang, Y. Zhou, D. D. Feng, and M. Chen, “Medical image classification with convolutional neural network,” in 2014 13th international conference on control automation robotics & vision (ICARCV).   IEEE, 2014, pp. 844–848.
  2. A. Mortazi and U. Bagci, “Automatically designing cnn architectures for medical image segmentation,” in Machine Learning in Medical Imaging: 9th International Workshop, MLMI 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 16, 2018, Proceedings 9.   Springer, 2018, pp. 98–106.
  3. F. Sultana, A. Sufian, and P. Dutta, “A review of object detection models based on convolutional neural network,” Intelligent computing: image processing based applications, pp. 1–16, 2020.
  4. X. Ma, Y. Niu, L. Gu, Y. Wang, Y. Zhao, J. Bailey, and F. Lu, “Understanding adversarial attacks on deep learning based medical image analysis systems,” Pattern Recognition, vol. 110, p. 107332, 2021.
  5. D. Wang, Z. Wang, L. Chen, H. Xiao, and B. Yang, “Cross-parallel transformer: Parallel vit for medical image segmentation,” Sensors, vol. 23, no. 23, p. 9488, 2023.
  6. Z. Li, Y. Li, Q. Li, P. Wang, D. Guo, L. Lu, D. Jin, Y. Zhang, and Q. Hong, “Lvit: language meets vision transformer in medical image segmentation,” IEEE transactions on medical imaging, 2023.
  7. F. Almalik, M. Yaqub, and K. Nandakumar, “Self-ensembling vision transformer (sevit) for robust medical image classification,” in International Conference on Medical Image Computing and Computer-Assisted Intervention.   Springer, 2022, pp. 376–386.
  8. S. A. Fezza, Y. Bakhti, W. Hamidouche, and O. Déforges, “Perceptual evaluation of adversarial attacks for cnn-based image classification,” in 2019 Eleventh International Conference on Quality of Multimedia Experience (QoMEX).   IEEE, 2019, pp. 1–6.
  9. K. Mahmood, R. Mahmood, and M. Van Dijk, “On the robustness of vision transformers to adversarial examples,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 7838–7847.
  10. I. J. Goodfellow, J. Shlens, and C. Szegedy, “Explaining and harnessing adversarial examples,” arXiv preprint arXiv:1412.6572, 2014.
  11. A. Mądry, A. Makelov, L. Schmidt, D. Tsipras, and A. Vladu, “Towards deep learning models resistant to adversarial attacks,” stat, vol. 1050, p. 9, 2017.
  12. Y. Lin, H. Zhao, X. Ma, Y. Tu, and M. Wang, “Adversarial attacks in modulation recognition with convolutional neural networks,” IEEE Transactions on Reliability, vol. 70, no. 1, pp. 389–401, 2021.
  13. R. Shao, Z. Shi, J. Yi, P.-Y. Chen, and C.-J. Hsieh, “On the adversarial robustness of vision transformers,” 2022.
  14. H. Qiu, Y. Zeng, Q. Zheng, S. Guo, T. Zhang, and H. Li, “An efficient preprocessing-based approach to mitigate advanced adversarial attacks,” IEEE Transactions on Computers, pp. 1–1, 2021.
  15. Y. Chang, H. Zhao, and W. Wang, “Enhancing the robustness of vision transformer defense against adversarial attacks based on squeeze-and-excitation module,” PeerJ Computer Science, vol. 9, p. e1197, 2023.
  16. S. G. Finlayson, H. W. Chung, I. S. Kohane, and A. L. Beam, “Adversarial attacks against medical deep learning systems,” arXiv preprint arXiv:1804.05296, 2018.
  17. M. Paschali, S. Conjeti, F. Navarro, and N. Navab, “Generalizability vs. robustness: investigating medical imaging networks using adversarial examples,” in Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part I.   Springer, 2018, pp. 493–501.
  18. S. Kaviani, K. J. Han, and I. Sohn, “Adversarial attacks and defenses on ai in medical imaging informatics: A survey,” Expert Systems with Applications, vol. 198, p. 116815, 2022.
  19. G. K. Dziugaite, Z. Ghahramani, and D. M. Roy, “A study of the effect of jpg compression on adversarial images,” arXiv preprint arXiv:1608.00853, 2016.
  20. A. Graese, A. Rozsa, and T. E. Boult, “Assessing threat of adversarial examples on deep neural networks,” in 2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA).   IEEE, 2016, pp. 69–74.
  21. W. Xu, D. Evans, and Y. Qi, “Feature squeezing: Detecting adversarial examples in deep neural networks,” in Proceedings 2018 Network and Distributed System Security Symposium, ser. NDSS 2018.   Internet Society, 2018. [Online]. Available: http://dx.doi.org/10.14722/ndss.2018.23198
  22. C. Guo, M. Rana, M. Cisse, and L. Van Der Maaten, “Countering adversarial images using input transformations,” arXiv preprint arXiv:1711.00117, 2017.
  23. A. Aldahdooh, W. Hamidouche, and O. Deforges, “Reveal of vision transformers robustness against adversarial attacks,” arXiv preprint arXiv:2106.03734, 2021.
  24. Y. Mo, D. Wu, Y. Wang, Y. Guo, and Y. Wang, “When adversarial training meets vision transformers: Recipes from training to architecture,” Advances in Neural Information Processing Systems, vol. 35, pp. 18 599–18 611, 2022.
  25. Y. Chang, H. Zhao, and W. Wang, “Ask-vit: A model with improved vit robustness through incorporating sk modules using adversarial training,” Electronics, vol. 11, no. 20, 2022. [Online]. Available: https://www.mdpi.com/2079-9292/11/20/3370
  26. R. Imam, M. Huzaifa, and M. E.-A. Azz, “On enhancing the robustness of vision transformers: Defensive diffusion,” 2023.
  27. O. N. Manzari, H. Ahmadabadi, H. Kashiani, S. B. Shokouhi, and A. Ayatollahi, “Medvit: A robust vision transformer for generalized medical image classification,” Computers in Biology and Medicine, vol. 157, p. 106791, 2023. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0010482523002561
  28. S. Jaeger, S. Candemir, S. Antani, Y.-X. J. Wáng, P.-X. Lu, and G. Thoma, “Two public chest x-ray datasets for computer-aided screening of pulmonary diseases,” Quantitative imaging in medicine and surgery, vol. 4, no. 6, p. 475, 2014.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com