Causal Explanations in Deep Learning Systems

No Thumbnail Available

Date

2025-06

Journal Title

Journal ISSN

Volume Title

Publisher

Indian Statistical Institute, Kolkata

Abstract

Deep learning models often deliver high predictive accuracy; however, their lack of interpretability can hinder their adoption in critical fields such as healthcare and finance. This thesis explores the concept of Intrinsic Causal Contribution (ICC), a novel method for explaining neural network predictions by quantifying each input feature’s intrinsic causal influence on the output, independent of correlated effects. ICC models the network as a Structural Causal Model and employs Causal Normalizing Flows to handle complex dependencies, with efficient estimation via the Jansen Estimator. Analysis on both synthetic and real data sets provides evidence that ICC produces faithful, interpretable attributions, often outperforming traditional approaches like SHAP and LIME. By revealing truly influential features, ICC supports transparent and responsible AI, especially in sensitive settings such as medical diagnosis.

Description

Dissertation under the supervision of Prof. Utpal Garain

Keywords

Intrinsic Causal Contribution (ICC),, SHAP, LIME, Deep Learning Systems

Citation

37p.

Endorsement

Review

Supplemented By

Referenced By