Are automated static analysis tools worth it? An investigation into relative warning density and external software quality on the example of Apache open source projects

2023 | journal article. A publication with affiliation to the University of Göttingen.

Jump to: Cite & Linked | Documents & Media | Details | Version history

Cite this publication

​Are automated static analysis tools worth it? An investigation into relative warning density and external software quality on the example of Apache open source projects​
Trautsch, A.; Herbold, S. & Grabowski, J.​ (2023) 
Empirical Software Engineering28(3).​ DOI: https://doi.org/10.1007/s10664-023-10301-2 

Documents & Media

document.pdf986.77 kBAdobe PDF

License

GRO License GRO License

Details

Authors
Trautsch, Alexander; Herbold, Steffen; Grabowski, Jens
Abstract
Abstract Automated Static Analysis Tools (ASATs) are part of software development best practices. ASATs are able to warn developers about potential problems in the code. On the one hand, ASATs are based on best practices so there should be a noticeable effect on software quality. On the other hand, ASATs suffer from false positive warnings, which developers have to inspect and then ignore or mark as invalid. In this article, we ask whether ASATs have a measurable impact on external software quality, using the example of PMD for Java. We investigate the relationship between ASAT warnings emitted by PMD on defects per change and per file. Our case study includes data for the history of each file as well as the differences between changed files and the project in which they are contained. We investigate whether files that induce a defect have more static analysis warnings than the rest of the project. Moreover, we investigate the impact of two different sets of ASAT rules. We find that, bug inducing files contain less static analysis warnings than other files of the project at that point in time. However, this can be explained by the overall decreasing warning density. When compared with all other changes, we find a statistically significant difference in one metric for all rules and two metrics for a subset of rules. However, the effect size is negligible in all cases, showing that the actual difference in warning density between bug inducing changes and other changes is small at best.
Issue Date
2023
Journal
Empirical Software Engineering 
ISSN
1382-3256
eISSN
1573-7616
Language
English
Sponsor
Deutsche Forschungsgemeinschaft https://doi.org/10.13039/501100001659
Universität Passau 100016135

Reference

Citations


Social Media