2000 character limit reached
XAI-Enhanced Semantic Segmentation Models for Visual Quality Inspection (2401.09900v1)
Published 18 Jan 2024 in cs.CV and cs.AI
Abstract: Visual quality inspection systems, crucial in sectors like manufacturing and logistics, employ computer vision and machine learning for precise, rapid defect detection. However, their unexplained nature can hinder trust, error identification, and system improvement. This paper presents a framework to bolster visual quality inspection by using CAM-based explanations to refine semantic segmentation models. Our approach consists of 1) Model Training, 2) XAI-based Model Explanation, 3) XAI Evaluation, and 4) Annotation Augmentation for Model Enhancement, informed by explanations and expert insights. Evaluations show XAI-enhanced models surpass original DeepLabv3-ResNet101 models, especially in intricate object segmentation.
- G. Baryannis, S. Dani, and G. Antoniou, “Predicting supply chain risks using machine learning: The trade-off between performance and interpretability,” Future Generation Computer Systems, vol. 101, pp. 993–1004, 2019.
- T. T. H. Nguyen, V. B. Truong, V. T. K. Nguyen, Q. H. Cao, and Q. K. Nguyen, “Towards trust of explainable ai in thyroid nodule diagnosis,” arXiv preprint arXiv:2303.04731, 2023.
- T. Clement, N. Kemmerzell, M. Abdelaal, and M. Amberg, “Xair: A systematic metareview of explainable ai (xai) aligned to the software development process,” Machine Learning and Knowledge Extraction, vol. 5, no. 1, pp. 78–108, 2023.
- X. Sun, J. Gu, S. Tang, and J. Li, “Research progress of visual inspection technology of steel products—a review,” Applied Sciences, vol. 8, no. 11, p. 2195, 2018.
- A. Q. Md, K. Jha, S. Haneef, A. K. Sivaraman, and K. F. Tee, “A review on data-driven quality prediction in the production process with machine learning for industry 4.0,” Processes, vol. 10, no. 10, p. 1966, 2022.
- Y. D. Yasuda, F. A. Cappabianco, L. E. G. Martins, and J. A. Gripp, “Aircraft visual inspection: A systematic literature review,” Computers in Industry, vol. 141, p. 103695, 2022.
- J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 779–788.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- S. Sundaram and A. Zeid, “Artificial intelligence-based smart quality inspection for manufacturing,” Micromachines, vol. 14, no. 3, p. 570, 2023.
- Y. Guo, Y. Liu, T. Georgiou, and M. S. Lew, “A review of semantic segmentation using deep neural networks,” International journal of multimedia information retrieval, vol. 7, pp. 87–93, 2018.
- J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440.
- L.-C. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinking atrous convolution for semantic image segmentation,” arXiv:1706.05587, 2017.
- S.-A. Rebuffi, R. Fong, X. Ji, and A. Vedaldi, “There and back again: Revisiting backpropagation saliency methods,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 8839–8848.
- A. Hedström, L. Weber, D. Bareeva, F. Motzkus, W. Samek, S. Lapuschkin, and M. M.-C. Höhne, “Quantus: an explainable ai toolkit for responsible evaluation of neural network explanations,” arXiv preprint arXiv:2202.06861, 2022.
- L. Weber, S. Lapuschkin, A. Binder, and W. Samek, “Beyond explaining: Opportunities and challenges of xai-based model improvement,” Information Fusion, 2022.
- S. A. Bargal, A. Zunino, V. Petsiuk, J. Zhang, K. Saenko, V. Murino, and S. Sclaroff, “Guided zoom: Questioning network evidence for fine-grained classification,” arXiv preprint arXiv:1812.02626, 2018.
- D. Schiller, T. Huber, F. Lingenfelser, M. Dietz, A. Seiderer, and E. André, “Relevance-based feature masking: Improving neural network based whale classification through explainable artificial intelligence,” 2019.
- H. Fukui, T. Hirakawa, T. Yamashita, and H. Fujiyoshi, “Attention branch network: Learning of attention mechanism for visual explanation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 10 705–10 714.
- J. ha Lee, I. hee Shin, S. gu Jeong, S.-I. Lee, M. Z. Zaheer, and B.-S. Seo, “Improvement in deep networks for optimization using explainable artificial intelligence,” in 2019 International Conference on Information and Communication Technology Convergence (ICTC). IEEE, 2019, pp. 525–530.
- J. M. Rožanec, L. Bizjak, E. Trajkova, P. Zajec, J. Keizer, B. Fortuna, and D. Mladenić, “Active learning and approximate model calibration for automated visual inspection in manufacturing,” arXiv preprint arXiv:2209.05486, 2022.
- R. Abdelfattah, X. Wang, and S. Wang, “Ttpla: An aerial-image dataset for detection and segmentation of transmission towers and power lines,” in Proceedings of the Asian Conference on Computer Vision, 2020.
- H. Wang, Z. Wang, M. Du, F. Yang, Z. Zhang, S. Ding, P. Mardziel, and X. Hu, “Score-cam: Score-weighted visual explanations for convolutional neural networks,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, 2020, pp. 24–25.
- C.-H. Chang, E. Creager, A. Goldenberg, and D. Duvenaud, “Explaining image classifiers by counterfactual generation,” arXiv preprint arXiv:1807.08024, 2018.
- R. Fu, Q. Hu, X. Dong, Y. Guo, Y. Gao, and B. Li, “Axiom-based grad-cam: Towards accurate visualization and explanation of cnns,” arXiv preprint arXiv:2008.02312, 2020.
- V. Petsiuk, R. Jain, V. Manjunatha, V. I. Morariu, A. Mehra, V. Ordonez, and K. Saenko, “Black-box explanation of object detectors via saliency maps,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 11 443–11 452.