Trustworthy explainable AI for optical network automation
Develop trustworthy explainable AI techniques that provide sufficient transparency of black-box AI models used for optical network automation so that operators can understand, trust, and reliably govern AI-driven decisions.
References
Further, according to , key AI research challenges remain open: ($i$) Training: Lack of available training datasets from real-world network deployments, ($ii$) Learning: Lack of lifelong (i.e., continual) learning, including AI degradation detection and model adaptation to progressive distribution shift, and ($iii$) Explainability: Lack of trustworthy explainable AI (XAI) due to insufficient transparency of blackbox AI.
— From Artificial Intelligence to Active Inference: The Key to True AI and 6G World Brain [Invited]
(2505.10569 - Maier, 29 Apr 2025) in Section 1 (Introduction), Point Alpha — Active Inference in Optical Networks