Software analytics has been the subject of considerable recent attention but is yet to receive significant industry traction. One of the key reasons is that software practitioners are reluctant to trust predictions produced by the analytics machinery without understanding the rationale for those predictions. While complex models such as deep learning and ensemble methods improve predictive performance, they have limited explainability. In this paper, we argue that making software analytics models explainable to software practitioners is as important as achieving accurate predictions. Explainability should therefore be a key measure for evaluating software analytics models.We envision that explainability will be a key driver for developing software analytics models that are useful in practice. We outline a research roadmap for this space, building on social science, explainable artificial intelligence and software engineering.
History
Pagination
53-56
Location
Gothenburg, Sweden
Start date
2018-05-27
End date
2018-06-03
ISSN
0270-5257
ISBN-13
9781450356626
Language
eng
Publication classification
E1 Full written paper - refereed
Copyright notice
2018, Association for Computing Machinery
Title of proceedings
ICSE-NIER 2018 : Proceedings of the 40th International Conference on Software Engineering: New Ideas and Emerging Results
Event
Software Engineering. International Conference (40th : 2018 : Gothenburg, Sweden)