Propagating Transparency: A Deep Dive into the Interpretability of Neural Networks
Permanent link
https://hdl.handle.net/10037/35446Date
2024-08-19Type
Journal articleTidsskriftartikkel
Peer reviewed
Abstract
In the rapidly evolving landscape of deep learning (DL), understanding the inner workings of neural networks remains a significant challenge. The need for transparency and accountability in DL models grows in importance as they become more prevalent in decision-making processes. Interpreting these models is key to addressing this challenge. This paper offers a comprehensive overview of interpretable methods for neural networks, particularly convolutional nets. The focus is on gradient-based propagation techniques that provide insight into the intricate mechanisms behind neural network predictions. Using a systematic review, we classify interpretability approaches that are based on gradients, dive into the theory of notable methods, and compare their strengths and weaknesses. Furthermore, we investigate different evaluation metrics for interpretable systems, often generalized under the term eXplainable Artificial Intelligence (XAI). We highlight the importance of these factors in evaluating the faithfulness, robustness, localization, complexity, randomization, and adherence to the axiomatic principles of XAI methods. Our objective is to assist researchers and practitioners in advancing towards a future for artificial intelligence that is characterized by a deeper understanding of its workings, thereby providing the desired transparency and accuracy. To this end, we offer a comprehensive summary of the latest advances in the field.
Publisher
Universitetet i OsloCitation
Somani A, Horsch A, Bopardikar, Prasad DK. Propagating Transparency: A Deep Dive into the Interpretability of Neural Networks. Nordic Machine Intelligence (NMI). 2024;4(2):1-18Metadata
Show full item recordCollections
Copyright 2024 The Author(s)