TY - JOUR
T1 - A Systematic Literature Review on Explainability for ML/DL-based Software Engineering
AU - Cao, Sicong
AU - Sun, Xiaobing
AU - Widyasari, Ratnadira
AU - Lo, David
AU - Wu, Xiaoxue
AU - Bo, Lili
AU - Zhang, Jiale
AU - Li, Bin
AU - Liu, Wei
AU - Wu, Di
AU - Chen, Yixin
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s). Publication rights licensed to ACM.
PY - 2025/10/25
Y1 - 2025/10/25
N2 - The remarkable achievements of Artificial Intelligence (AI) algorithms, particularly in Machine Learning (ML) and Deep Learning (DL), have fueled their extensive deployment across multiple sectors, including Software Engineering (SE). However, due to their black-box nature, these promising AI-driven SE models are still far from being deployed in practice. This lack of explainability poses unwanted risks for their applications in critical tasks, such as vulnerability detection, where decision-making transparency is of paramount importance. This article endeavors to elucidate this interdisciplinary domain by presenting a systematic literature review of approaches that aim to improve the explainability of AI models within the context of SE. The review canvasses work appearing in the most prominent SE and AI conferences and journals, and spans 108 articles across 23 unique SE tasks. Based on three key Research Questions (RQs), we aim to (1) summarize the SE tasks where XAI techniques have shown success to date; (2) classify and analyze different XAI techniques; and (3) investigate existing evaluation approaches. Based on our findings, we identified a set of challenges remaining to be addressed in existing studies, together with a set of guidelines highlighting potential opportunities we deemed appropriate and important for future work.
AB - The remarkable achievements of Artificial Intelligence (AI) algorithms, particularly in Machine Learning (ML) and Deep Learning (DL), have fueled their extensive deployment across multiple sectors, including Software Engineering (SE). However, due to their black-box nature, these promising AI-driven SE models are still far from being deployed in practice. This lack of explainability poses unwanted risks for their applications in critical tasks, such as vulnerability detection, where decision-making transparency is of paramount importance. This article endeavors to elucidate this interdisciplinary domain by presenting a systematic literature review of approaches that aim to improve the explainability of AI models within the context of SE. The review canvasses work appearing in the most prominent SE and AI conferences and journals, and spans 108 articles across 23 unique SE tasks. Based on three key Research Questions (RQs), we aim to (1) summarize the SE tasks where XAI techniques have shown success to date; (2) classify and analyze different XAI techniques; and (3) investigate existing evaluation approaches. Based on our findings, we identified a set of challenges remaining to be addressed in existing studies, together with a set of guidelines highlighting potential opportunities we deemed appropriate and important for future work.
KW - Explainable AI
KW - interpretability
KW - neural networks
KW - survey
KW - XAI
UR - https://www.scopus.com/pages/publications/105023820576
U2 - 10.1145/3763230
DO - 10.1145/3763230
M3 - Article
AN - SCOPUS:105023820576
SN - 0360-0300
VL - 58
JO - ACM Computing Surveys
JF - ACM Computing Surveys
IS - 4
M1 - 95
ER -