A Survey on Explainable Artificial Intelligence (XAI) in Deep Learning Systems

Authors

  • Dr. Ankit Jagga Assistant Professor, Computer Science, Sri Guru Teg Bahadur Khalsa College for Girls, Aakar, Patiala

DOI:

https://doi.org/10.31305/rrijm.2024.v09.n06.047

Keywords:

Explainable Artificial Intelligence, Deep Learning, Interpretability, SHAP, LIME, Grad-CAM, Model Transparency, Trustworthy AI

Abstract

Artificial Intelligence has rapidly evolved over the past decade, with deep learning models achieving remarkable success in complex tasks such as medical diagnosis, natural language processing, financial prediction, and autonomous navigation. Despite their high performance, these models often operate as opaque systems whose internal decision-making processes are difficult for humans to understand. This lack of transparency has led to concerns regarding trust, accountability, bias, and ethical compliance. Explainable Artificial Intelligence (XAI) has emerged as a critical research area aimed at addressing these limitations by providing interpretability and insight into model behavior. This survey explores the foundations of XAI, categorizes existing explanation methods, reviews significant literature contributions, discusses practical applications, highlights current challenges, and outlines future research directions. The paper emphasizes the importance of XAI in enabling reliable and responsible AI deployment in safety-critical and socially sensitive domains.

References

Arrieta, A. B., Díaz-Rodríguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., García, S., Gil-López, S., Molina, D., Benjamins, R., Chatila, R., & Herrera, F. (2020). Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion, 58, 82–115.

Doshi-Velez, F., & Kim, B. (2017). Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608.

Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., & Pedreschi, D. (2018). A survey of methods for explaining black box models. ACM Computing Surveys, 51(5), 1–42.

Lundberg, S. M., & Lee, S. I. (2017). A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NeurIPS) (pp. 4768–4777).

Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 1135–1144).

Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., & Batra, D. (2017). Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV) (pp. 618–626).

Samek, W., Wiegand, T., & Müller, K. R. (2017). Explainable artificial intelligence: Understanding, visualizing and interpreting deep learning models. arXiv preprint arXiv:1708.08296.

Montavon, G., Samek, W., & Müller, K. R. (2018). Methods for interpreting and understanding deep neural networks. Digital Signal Processing, 73, 1–15.

Rudin, C. (2019). Stop explaining black box machine learning models for high-stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5), 206–215.

Adadi, A., & Berrada, M. (2018). Peeking inside the black-box: A survey on explainable artificial intelligence (XAI). IEEE Access, 6, 52138–52160.

Zhang, Q., & Zhu, S. C. (2018). Visual interpretability for deep learning: A survey. Frontiers of Information Technology & Electronic Engineering, 19(1), 27–39.

Molnar, C. (2022). Interpretable Machine Learning (2nd ed.). Lulu.com.

Downloads

Published

14-06-2024

How to Cite

Jagga, A. (2024). A Survey on Explainable Artificial Intelligence (XAI) in Deep Learning Systems . RESEARCH REVIEW International Journal of Multidisciplinary, 9(6), 372–376. https://doi.org/10.31305/rrijm.2024.v09.n06.047