Empirical evaluation of technical terms in open source project issues

  • Joselito Mota Júnior UFBA
  • Ivan Machado UFBA

Abstract


The growing demand for robust and reliable software systems has led the scientific community and the software industry to investigate aspects related to defects reported in bug tracking systems. The overall goal is to understand the causes of defects, and thus establish strategies to prevent their occurrence and spread, reducing possible damage. Defects reported by users via issues in bug trackers reveal important information for developers. The present investigation carried out an analysis of data from 510,212 issues from 85 repositories of open source projects, available on Github. With the help of a list of 608 technical terms selected manually, the frequency in the use of terms by contributors to these repositories was analyzed. The study found the use of terms in the different fields of an issue, as well as the possibility of identifying the terms most used in these repositories. The results of this study point to potential gains in understanding the failure environment by presenting terms that are indicative of defects in fundamental repository structures.

Keywords: Software Engineering, data mining, defects, issue, technical terms

References

Abal, I., Brabrand, C., and Wasowski, A. (2014). 42 variability bugs in the linux kernel: a qualitative analysis. In Proceedings of the 29th ACM/IEEE International Conference on Automated Software Engineering (ASE), pages 421–432.

Boechat, G., Junior, J. M., Machado, I., and Mendonca, M. (2019). Analise de sentimentos em discussoes de issues reabertas do github. In Anais do VII Workshop on Software Visualization, Evolution and Maintenance (VEM), pages 13–20. SBC.

Card, D. N. (1998). Learning from our mistakes with defect causal analysis. IEEE software, 15(1):56–63.

Catolino, G., Palomba, F., Zaidman, A., and Ferrucci, F. (2019). Not all bugs are the same: Understanding, characterizing, and classifying bug types. Journal of Systems and Software, 152:165–181.

Grady, R. B. (1996). Software failure analysis for high-return process improvement decisions. Hewlett Packard Journal, 47:15–24.

Gromova, A., Itkin, I., Pavlov, S., and Korovayev, A. (2019). Raising the quality of bug reports by predicting software defect indicators. In 2019 IEEE 19th International Conference on Software Quality, Reliability and Security Companion (QRS-C), pages 198–204. IEEE.

Karahroudy, A. A. and Tabrizi, M. (2011). Software defect taxonomy, analysis and overview. In Proceedings of the International Conference on Software Engineering Research and Practice (SERP).

Sullivan, M. and Chillarege, R. (1991). Software Defects and their Impact on System Availability: A Study of Field Failures in Operating Systems. In International Symposium on Fault-Tolerant Computing, volume 21, pages 2–9.

Zubrow, D. (2009). IEEE Standard Classification for Software Anomalies. IEEE Computer Society.
Published
2020-06-30
MOTA JÚNIOR, Joselito ; MACHADO, Ivan. Empirical evaluation of technical terms in open source project issues. In: SBC UNDERGRADUATE RESEARCH CONTEST (CTIC-SBC), 39. , 2020, Cuiabá. Anais [...]. Porto Alegre: Sociedade Brasileira de Computação, 2020 . p. 11-20.