Interpretable Deep Learning Models: Enhancing Transparency and Trustworthiness in Explainable AI

Main Article Content

R. S. Deshpande, P. V. Ambatkar

Abstract

Explainable AI (XAI) aims to address the opacity of deep learning models, which can limit their adoption in critical decision-making applications. This paper presents a novel framework that integrates interpretable components and visualization techniques to enhance the transparency and trustworthiness of deep learning models. We propose a hybrid explanation method combining saliency maps, feature attribution, and local interpretable model-agnostic explanations (LIME) to provide comprehensive insights into the model's decision-making process.


Our experiments with convolutional neural networks (CNNs) and transformers demonstrate that our approach improves interpretability without compromising performance. User studies with domain experts indicate that our visualization dashboard facilitates better understanding and trust in AI systems. This research contributes to developing more transparent and trustworthy deep learning models, paving the way for broader adoption in sensitive applications where human users need to understand and trust AI decisions.

Article Details

Section
Articles