Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Measuring the Quality of Explanations: The System Causability Scale (SCS). Comparing Human and Machine Explanations (1912.09024v1)

Published 19 Dec 2019 in cs.AI
Measuring the Quality of Explanations: The System Causability Scale (SCS). Comparing Human and Machine Explanations

Abstract: Recent success in AI and Machine Learning (ML) allow problem solving automatically without any human intervention. Autonomous approaches can be very convenient. However, in certain domains, e.g., in the medical domain, it is necessary to enable a domain expert to understand, why an algorithm came up with a certain result. Consequently, the field of Explainable AI (xAI) rapidly gained interest worldwide in various domains, particularly in medicine. Explainable AI studies transparency and traceability of opaque AI/ML and there are already a huge variety of methods. For example with layer-wise relevance propagation relevant parts of inputs to, and representations in, a neural network which caused a result, can be highlighted. This is a first important step to ensure that end users, e.g., medical professionals, assume responsibility for decision making with AI/ML and of interest to professionals and regulators. Interactive ML adds the component of human expertise to AI/ML processes by enabling them to re-enact and retrace AI/ML results, e.g. let them check it for plausibility. This requires new human-AI interfaces for explainable AI. In order to build effective and efficient interactive human-AI interfaces we have to deal with the question of how to evaluate the quality of explanations given by an explainable AI system. In this paper we introduce our System Causability Scale (SCS) to measure the quality of explanations. It is based on our notion of Causability (Holzinger et al., 2019) combined with concepts adapted from a widely accepted usability scale.

An Analysis of the System Causability Scale (SCS)

The paper "Measuring the Quality of Explanations: The System Causability Scale (SCS)" by Andreas Holzinger, Andre Carrington, and Heimo Müller focuses on an emergent area in the domain of Explainable Artificial Intelligence (xAI). The authors introduce the System Causability Scale (SCS) as a metric for evaluating the quality of explanations produced by AI systems, emphasizing the importance of human-AI interface reliability.

Core Proposition and Context

The increasing use of AI, particularly in sensitive domains such as medicine, has heightened the demand for transparency and interpretability. The opaque nature of many machine learning models, especially those based on deep learning, presents challenges in trust and accountability. Consequently, xAI has become central to ensuring that such systems provide comprehensible and justifiable outputs.

The authors argue that while existing usability metrics, such as the System Usability Scale (SUS), evaluate user interaction with systems, there is a lack of standardized methodologies for measuring the effectiveness and causality of explanations provided by AI systems. The SCS is proposed to fill this gap, offering a framework that assesses the extent to which AI explanations facilitate understanding and convey causality efficiently and effectively.

Construction of SCS

The SCS is an adaptation of the widely recognized SUS. It is structured to evaluate AI system explanations through a questionnaire using the Likert scale. Participants score explanations based on their comprehensibility, relevance, and the extent to which they aid in discerning causality. Unlike SUS, SCS explicitly focuses on the causal aspects of explanations, forming a bridge between AI outputs and user interpretation.

Implications and Findings

The utility of SCS extends beyond theoretical applications. The paper highlights its application in the medical domain, notably using the Framingham Risk Tool as a test case. The case paper demonstrates SCS's ability to gauge the quality of output explanations in a real-world context. The findings suggest that SCS can reliably measure end-user satisfaction and the understandability of AI explanations, fostering better integration of AI into human decision-making processes.

Results presented in the paper underline that explanations should be delivered in a manner that is contextually relevant and aligned with the user’s existing knowledge base. The necessity for explanations to enable users to adjust the level of detail, comprehend information with minimal external aid, and encounter no inconsistencies is also emphasized.

Future Directions and Challenges

The introduction of the SCS provides a valuable tool for the AI research community, aiding in the development of AI systems that prioritize transparency and user-centric designs. However, the paper recognizes limitations inherent in Likert scale methodologies, acknowledging the potential for disparities in perceived interval scales among respondents.

Future work as envisioned by the authors includes extensive empirical evaluations to refine the SCS and ensure its applicability across various domains beyond healthcare. Development of methods to quantify and communicate uncertainty in AI explanations could further cement the role of SCS in xAI. Broader application of the SCS promises to enhance how interactive machine learning (iML) systems are evaluated, with implications for the design of algorithms that are not only technically robust but also intuitively understandable to end-users.

In conclusion, the System Causability Scale emerges as a pivotal contribution to the field of xAI, enabling comprehensive assessment of AI systems’ explainability. The notion that causability should be directly measured emphasizes the need for AI systems to support human understanding and decision-making effectively. As AI systems become deeply integrated across various sectors, tools like the SCS will be critical in ensuring these systems operate transparently and align with human cognitive processes.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Andreas Holzinger (26 papers)
  2. André Carrington (1 paper)
  3. Heimo Müller (9 papers)
Citations (276)