Dynamic Importance Learning using Fisher Information Matrix (FIM) for Nonlinear Dynamic Mapping (2406.05395v2)
Abstract: Understanding output variance is critical in modeling nonlinear dynamic systems, as it reflects the system's sensitivity to input variations and feature interactions. This work presents a methodology for dynamically determining relevance scores in black-box models while ensuring interpretability through an embedded decision module. This interpretable module, integrated into the first layer of the model, employs the Fisher Information Matrix (FIM) and logistic regression to compute relevance scores, interpreted as the probabilities of input neurons being active based on their contribution to the output variance. The proposed method leverages a gradient-based framework to uncover the importance of variance-driven features, capturing both individual contributions and complex feature interactions. These relevance scores are applied through element-wise transformations of the inputs, enabling the black-box model to prioritize features dynamically based on their impact on system output. This approach effectively bridges interpretability with the intricate modeling of nonlinear dynamics and time-dependent interactions. Simulation results demonstrate the method's ability to infer feature interactions while achieving superior performance in feature relevance compared to existing techniques. The practical utility of this approach is showcased through its application to an industrial pH neutralization process, where critical system dynamics are uncovered.