Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Functional trustworthiness of AI systems by statistically valid testing (2310.02727v1)

Published 4 Oct 2023 in stat.ML, cs.AI, and cs.LG

Abstract: The authors are concerned about the safety, health, and rights of the European citizens due to inadequate measures and procedures required by the current draft of the EU AI Act for the conformity assessment of AI systems. We observe that not only the current draft of the EU AI Act, but also the accompanying standardization efforts in CEN/CENELEC, have resorted to the position that real functional guarantees of AI systems supposedly would be unrealistic and too complex anyways. Yet enacting a conformity assessment procedure that creates the false illusion of trust in insufficiently assessed AI systems is at best naive and at worst grossly negligent. The EU AI Act thus misses the point of ensuring quality by functional trustworthiness and correctly attributing responsibilities. The trustworthiness of an AI decision system lies first and foremost in the correct statistical testing on randomly selected samples and in the precision of the definition of the application domain, which enables drawing samples in the first place. We will subsequently call this testable quality functional trustworthiness. It includes a design, development, and deployment that enables correct statistical testing of all relevant functions. We are firmly convinced and advocate that a reliable assessment of the statistical functional properties of an AI system has to be the indispensable, mandatory nucleus of the conformity assessment. In this paper, we describe the three necessary elements to establish a reliable functional trustworthiness, i.e., (1) the definition of the technical distribution of the application, (2) the risk-based minimum performance requirements, and (3) the statistically valid testing based on independent random samples.

Functional Trustworthiness of AI Systems: A Critical Examination

The reviewed paper presents a discernible evaluation of the current regulatory state and standardization efforts underpinning the European Union's AI Act, with a concentrated focus on the pivotal role of functional trustworthiness through statistically valid testing. The discourse champions the statistical methods fundamental to ML and deep learning (DL) as central tenets for assessing AI systems' robustness, accuracy, and transparency, arguing that existing regulations fall short in these aspects.

Key Arguments and Concepts

The paper emphasizes the intrinsic necessity of functional trustworthiness, asserting that AI systems must undergo empirically valid statistical tests on independent and random samples to ensure they meet predefined performance standards. The concept is encapsulated as consisting of three foundational elements:

  1. Definition of the Technical Distribution: Establishing a precise application domain is crucial. This involves characterizing the technical distribution, which allows for creating representative random samples imperative for testing. Such clarity ensures model performance is accurately measured against the identified domain.
  2. Risk-Based Minimum Performance Requirements: The paper positions risk analysis at the core of system development, advocating that performance metrics must emanate from a thorough understanding of the application's risks. This appraisal should guide defining acceptable operational thresholds encompassing safety and non-discrimination.
  3. Statistically Valid Testing: Testing AI models through randomly sampled data from the defined distribution is critical for assessing performance. This statistical approach ensures an AI system performs as intended within its deployment scope, thus addressing concerns about the unpredictability inherent in high-complexity models like those used in DL.

Criticisms of the Current EU AI Act

The authors critique the EU AI Act for emphasizing documentation over the empirical validation of AI system quality. They suggest the Act's framework potentially encourages insufficiently tested AI solutions to enter the market under a false guise of reliability. The inadequacy of provisions concerning random sampling and statistical validation in testing requirements is highly scrutinized, underscoring a gap in aligning regulatory guidelines with established ML principles.

Practical and Theoretical Implications

Practically, the paper outlines how adherence to these methodological principles could significantly enhance AI systems' trustworthiness and reliability, leading to better risk management in real-world applications. Such an approach supports a finer granularity in AI system deployment, emphasizing the need for application-specific testing to mitigate potential biases and vulnerabilities. Theoretically, the discussion solidifies the necessity of bridging traditional engineering approaches with data-driven techniques to forge robust AI solutions.

Future Directions

The paper advocates for a reformed AI regulatory environment that integrates functional trustworthiness as a core component of conformity assessment. It speculates that future developments may see more rigorous enforcement of statistical testing protocols, potentially leading to new industry standards harmonized globally.

Additionally, the discussion on AI systems, such as personal AI assistants and the trade-offs between creativity and ethical outputs, suggests an avenue for future empirical research to explore the balance between functionality and ethical constraints, especially as personal assistants become more entrenched in daily life.

Conclusion

This paper provides a comprehensive examination of the critical importance of statistically valid testing in establishing functional trustworthiness for AI systems. It makes a compelling case that existing regulatory frameworks, like the EU AI Act, are yet to fully integrate these essential principles into their core. By focusing on empirical validation through statistically driven methodologies and advocating for a harmonized approach in AI system regulation, the presented arguments pave crucial pathways for future policy enhancements and technological advancements in the field of artificial intelligence.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. Aws Albarghouthi et al. Introduction to neural network verification. Foundations and Trends® in Programming Languages, 7(1–2):1–157, 2021.
  2. Concrete problems in ai safety. arXiv preprint arXiv:1606.06565, 2016.
  3. Towards auditable ai systems: Current status and future directions. White paper, Federal Office for Information Security and University of Cagliari and TÜV Association and Others, May 2021. Based on the workshop "Auditing AI-Systems: From Basics to Applications", October 6th 2020, Fraunhofer Forum, Berlin.
  4. Towards auditable ai systems: From principles to practice. White paper, Federal Office for Information Security Germany and TÜV-Verband and Fraunhofer HHI and Others, May 2022. Based on the 2nd international Workshop "Towards Auditable AI Systems", October 26th 2021, Fraunhofer Forum Digitale Technologien, Berlin.
  5. Certified adversarial robustness via randomized smoothing. In international conference on machine learning, pages 1310–1320. PMLR, 2019.
  6. Impossibility theorems for domain adaptation. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pages 129–136. JMLR Workshop and Conference Proceedings, 2010.
  7. Pattern classification. John Wiley & Sons, 1973.
  8. COM(2018) 237 final. Communication from the commission to the european parliament, the european council, the council, the european economic and social committee and the committee of the regions: Artificial intelligence for europe, 2018. URL https://eur-lex.europa.eu/legal-content/EN/ALL/?uri=COM:2018:237:FIN. Online, accessed 2023-05-31.
  9. COM(2021) 206 final. Proposal for a regulation of the european parliament and of the council laying down harmonised rules on artificial intelligence (artificial intelligence act) and amending certain union legislative acts, 2021. URL https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=celex%3A52021PC0206. Online, accessed 2023-06-01.
  10. Ai system engineering—key challenges and lessons learned. Machine Learning and Knowledge Extraction, 3(1):56–83, 2020.
  11. Ai2: Safety and robustness certification of neural networks with abstract interpretation. In 2018 IEEE symposium on security and privacy (SP), pages 3–18. IEEE, 2018.
  12. Generative adversarial networks. arXiv preprint arXiv:1406.2661, 2014.
  13. Not what you’ve signed up for: Compromising real-world llm-integrated applications with indirect prompt injection, 2023.
  14. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840–6851, 2020.
  15. Provable lipschitz certification for generative models. In International Conference on Machine Learning, pages 5118–5126. PMLR, 2021.
  16. Leakage in data mining: Formulation, detection, and avoidance. ACM Transactions on Knowledge Discovery from Data (TKDD), 6(4):1–21, 2012.
  17. Openassistant conversations–democratizing large language model alignment. arXiv preprint arXiv:2304.07327, 2023.
  18. Trivial or impossible–dichotomous data difficulty masks model differences (on imagenet and beyond). CoRR, abs/2110.05922, 2021. URL https://arxiv.org/abs/2110.05922.
  19. P9_TA(2023)0236. Artificial intelligence act, 2021. URL https://www.europarl.europa.eu/doceo/document/TA-9-2023-0236_EN.html. Online, accessed 2023-09-26.
  20. Dataset shift in machine learning. Mit Press, 2008.
  21. Zero-shot text-to-image generation. In International Conference on Machine Learning, pages 8821–8831. PMLR, 2021.
  22. A generalist agent. arXiv preprint arXiv:2205.06175, 2022.
  23. A method for obtaining digital signatures and public-key cryptosystems. Communications of the ACM, 21(2):120–126, 1978.
  24. Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100, 2023.
  25. Introducing chatgpt, Nov 2022. URL https://openai.com/blog/chatgpt. Online, accessed 2023-05-17.
  26. The curse of recursion: Training on generated data makes models forget, 2023.
  27. Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199, 2013.
  28. Alpaca: A strong, replicable instruction-following model. Stanford Center for Research on Foundation Models. https://crfm. stanford. edu/2023/03/13/alpaca. html, 2023.
  29. Lamda: Language models for dialog applications. arXiv preprint arXiv:2201.08239, 2022.
  30. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  31. On the robustness of chatgpt: An adversarial and out-of-distribution perspective. arXiv preprint arXiv:2302.12095, 2023.
  32. Norbert Wiener. Some moral and technical consequences of automation: As machines learn they may develop unforeseen strategies at rates that baffle their programmers. Science, 131(3410):1355–1358, 1960.
  33. Trusted artificial intelligence: Towards certification of machine learning applications. arXiv preprint arXiv:2103.16910, 2021.
  34. On generalization in moment-based domain adaptation. Annals of Mathematics and Artificial Intelligence, 89:333–369, 2021.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Bernhard Nessler (6 papers)
  2. Thomas Doms (2 papers)
  3. Sepp Hochreiter (82 papers)
Youtube Logo Streamline Icon: https://streamlinehq.com