Explaining Neural Networks

Understanding the function learned by a neural network is crucial in many domains, e.g., to detect a model’s adaption to concept drift in online learning. Existing global surrogate model approaches generate explanations by maximizing the fidelity between the neural network and a surrogate model on a sample-basis. Our idea, instead, is to generate explanations via another neural network (which we call I-Net), which maps network parameters to a symbolic representation of the network function. The training of an I-Net for a family of functions can be performed up-front and subsequent generation of an explanation only requires querying the I-Net once, which is computationally very efficient and does not require training data.

Key Publications

Sascha Marton, Stefan Lüdtke, Christian Bartelt. Explanations for neural networks by neural networks. Applied Sciences 2022. [web]

Sascha Marton, Stefan Lüdtke , Christian Bartelt, Andrej Tschalzev, Heiner Stuckenschmidt. Explaining Neural Networks without Access to Training Data. arXiv preprint arXiv:2206.04891. 2022. [web]

Contact

Stefan Lüdtke

stefan.luedtke2@uni-rostock.de