Causal Explanations in Deep Learning Systems
No Thumbnail Available
Date
2025-06
Authors
Journal Title
Journal ISSN
Volume Title
Publisher
Indian Statistical Institute, Kolkata
Abstract
Deep learning models often deliver high predictive accuracy; however, their lack of interpretability
can hinder their adoption in critical fields such as healthcare and finance. This
thesis explores the concept of Intrinsic Causal Contribution (ICC), a novel method for explaining
neural network predictions by quantifying each input feature’s intrinsic causal influence on
the output, independent of correlated effects. ICC models the network as a Structural Causal
Model and employs Causal Normalizing Flows to handle complex dependencies, with efficient
estimation via the Jansen Estimator. Analysis on both synthetic and real data sets provides
evidence that ICC produces faithful, interpretable attributions, often outperforming traditional
approaches like SHAP and LIME. By revealing truly influential features, ICC supports transparent
and responsible AI, especially in sensitive settings such as medical diagnosis.
Description
Dissertation under the supervision of Prof. Utpal Garain
Keywords
Intrinsic Causal Contribution (ICC),, SHAP, LIME, Deep Learning Systems
Citation
37p.
