Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/2897695.2897696acmconferencesArticle/Chapter ViewAbstractPublication PagesicseConference Proceedingsconference-collections
research-article

Measuring high and low priority defects on traditional and mobile open source software

Published: 14 May 2016 Publication History

Abstract

Software defects are the major cause for system failures. To effectively design tools and provide support for detecting and recovering from software failures, requires a deep understanding of defect features. In this paper we present an analysis of defect characteristics in two different open source software development domains: Mobile and Traditional. Our attention is focused on measuring the differences between High-Priority and Low-Priority defects. High or Low priority of a given defect is decided by a developer when creating a bug report for an issue tracking system. We sampled hundreds of real world bugs in hundreds of large and representative open-source projects. We used natural language text classification techniques to automatically analyse roughly 700,000 bug reports from the Bugzilla, Jira and Google Issues issue tracking systems. Results show that there are differences between High-Priority and Low-Priority defects classification in Mobile and Traditional development domains.

References

[1]
C. Bird, A. Bachmann, E. Aune, J. Duffy, A. Bernstein, V. Filkov, and P. Devanbu. Fair and balanced?: bias in bug-fix datasets. In Proceedings of the the 7th joint meeting of the European software engineering conference and the ACM SIGSOFT symposium on The foundations of software engineering, pages 121--130. ACM, 2009.
[2]
C. Catal and B. Diri. A systematic review of software fault prediction studies. Expert systems with applications, 36(4):7346--7354, 2009.
[3]
G. Concas, G. Destefanis, M. Marchesi, M. Ortu, and R. Tonelli. Micro patterns in agile software. In Agile Processes in Software Engineering and Extreme Programming: 14th International Conference, XP 2013, Vienna, Austria, June 3-7, 2013, Proceedings, volume 149, page 210. Springer, 2013.
[4]
G. Destefanis, S. Counsell, G. Concas, and R. Tonelli. Software metrics in agile software: An empirical study. In Agile Processes in Software Engineering and Extreme Programming, pages 157--170. Springer, 2014.
[5]
G. Destefanis, R. Tonelli, G. Concas, and M. Marchesi. An analysis of anti-micro-patterns effects on fault-proneness in large java systems. In Proceedings of the 27th Annual ACM Symposium on Applied Computing, pages 1251--1253. ACM, 2012.
[6]
G. Destefanis, R. Tonelli, E. Tempero, G. Concas, and M. Marchesi. Micro pattern fault-proneness. In Software Engineering and Advanced Applications (SEAA), 2012 38th EUROMICRO Conference on, pages 302--306. IEEE, 2012.
[7]
S. Di Martino, F. Ferrucci, C. Gravino, and F. Sarro. A genetic algorithm to configure support vector machines for predicting fault-prone components. In Product-Focused Software Process Improvement, pages 247--261. Springer, 2011.
[8]
M. D'Ambros, M. Lanza, and R. Robbes. Evaluating defect prediction approaches: a benchmark and an extensive comparison. Empirical Software Engineering, 17(4--5):531--577, 2012.
[9]
A. Guzzi, A. Bacchelli, M. Lanza, M. Pinzger, and A. v. Deursen. Communication in open source software development mailing lists. In Proceedings of the Tenth International Workshop on Mining Software Repositories, pages 277--286. IEEE Press, 2013.
[10]
T. Gyimothy, R. Ferenc, and I. Siket. Empirical validation of object-oriented metrics on open source software for fault prediction. Software Engineering, IEEE Transactions on, 31(10):897--910, 2005.
[11]
T. Hall, S. Beecham, D. Bowes, D. Gray, and S. Counsell. A systematic literature review on fault prediction performance in software engineering. Software Engineering, IEEE Transactions on, 38(6):1276--1304, 2012.
[12]
T. Hall and D. Bowes. The state of machine learning methodology in software fault prediction. In Machine Learning and Applications (ICMLA), 2012 11th International Conference on, volume 2, pages 308--313. IEEE, 2012.
[13]
M. Harman, S. Islam, Y. Jia, L. L. Minku, F. Sarro, and K. Srivisut. Less is more: Temporal fault predictive performance over multiple hadoop releases. In Search-Based Software Engineering, pages 240--246. Springer, 2014.
[14]
A. E. Hassan. Predicting faults using the complexity of code changes. In Proceedings of the 31st International Conference on Software Engineering, pages 78--88. IEEE Computer Society, 2009.
[15]
K. Herzig, S. Just, and A. Zeller. It's not a bug, it's a feature: how misclassification impacts bug prediction. In Proceedings of the 2013 International Conference on Software Engineering, pages 392--401. IEEE Press, 2013.
[16]
T. Jiang, L. Tan, and S. Kim. Personalized defect prediction. In Automated Software Engineering (ASE), 2013 IEEE/ACM 28th International Conference on, pages 279--289. IEEE, 2013.
[17]
E. Kalliamvakou, G. Gousios, K. Blincoe, L. Singer, D. M. German, and D. Damian. The promises and perils of mining github. In Proceedings of the 11th Working Conference on Mining Software Repositories, pages 92--101. ACM, 2014.
[18]
S. Kim, E. J. Whitehead Jr, and Y. Zhang. Classifying software changes: Clean or buggy? Software Engineering, IEEE Transactions on, 34(2):181--196, 2008.
[19]
S. Kim, T. Zimmermann, E. J. Whitehead Jr, and A. Zeller. Predicting faults from cached history. In Proceedings of the 29th international conference on Software Engineering, pages 489--498. IEEE Computer Society, 2007.
[20]
Z. Li, L. Tan, X. Wang, S. Lu, Y. Zhou, and C. Zhai. Have things changed now?: an empirical study of bug characteristics in modern open source software. In Proceedings of the 1st workshop on Architectural and system support for improving software dependability, pages 25--33. ACM, 2006.
[21]
S. Matsumoto, Y. Kamei, A. Monden, K.-i. Matsumoto, and M. Nakamura. An analysis of developer metrics for fault prediction. In Proceedings of the 6th International Conference on Predictive Models in Software Engineering, page 18. ACM, 2010.
[22]
M. Ortu, G. Destefanis, M. Orru, R. Tonelli, and M. Marchesi. Could micro patterns be used as software stability indicator? In Patterns Promotion and Anti-patterns Prevention (PPAP), 2015 IEEE 2nd Workshop on, pages 11--12. IEEE, 2015.
[23]
F. Sarro, S. Di Martino, F. Ferrucci, and C. Gravino. A further analysis on the use of genetic algorithm to configure support vector machines for inter-release fault prediction. In Proceedings of the 27th annual ACM symposium on applied computing, pages 1215--1220. ACM, 2012.
[24]
M. Shepperd, D. Bowes, and T. Hall. Researcher bias: The use of machine learning in software defect prediction. Software Engineering, IEEE Transactions on, 40(6):603--616, 2014.
[25]
E. Shihab, A. Mockus, Y. Kamei, B. Adams, and A. E. Hassan. High-impact defects: a study of breakage and surprise defects. In Proceedings of the 19th ACM SIGSOFT symposium and the 13th European conference on Foundations of software engineering, pages 300--310. ACM, 2011.
[26]
Q. Song, Z. Jia, M. Shepperd, S. Ying, and S. Y. J. Liu. A general software defect-proneness prediction framework. Software Engineering, IEEE Transactions on, 37(3):356--370, 2011.
[27]
L. Tan, C. Liu, Z. Li, X. Wang, Y. Zhou, and C. Zhai. Bug characteristics in open source software. Empirical Software Engineering, pages 1--41, 2013.
[28]
Y. Zhou and H. Leung. Empirical analysis of object-oriented design metrics for predicting high and low severity faults. Software Engineering, IEEE Transactions on, 32(10):771--789, 2006.

Cited By

View all
  • (2022)MULA: A Just-In-Time Multi-labeling System for Issue ReportsIEEE Transactions on Reliability10.1109/TR.2021.307451271:1(250-263)Online publication date: Mar-2022
  • (2022)Detecting non-natural language artifacts for de-noising bug reportsAutomated Software Engineering10.1007/s10515-022-00350-029:2Online publication date: 1-Nov-2022
  • (2021)Just-in-time defect prediction for Android apps via imbalanced deep learning modelProceedings of the 36th Annual ACM Symposium on Applied Computing10.1145/3412841.3442019(1447-1454)Online publication date: 22-Mar-2021
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
WETSoM '16: Proceedings of the 7th International Workshop on Emerging Trends in Software Metrics
May 2016
76 pages
ISBN:9781450341776
DOI:10.1145/2897695
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 May 2016

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. bug categorisation
  2. bug reports
  3. data mining

Qualifiers

  • Research-article

Funding Sources

  • Engineering and Physical Sciences Research Council (EPSRC) of the UK

Conference

ICSE '16
Sponsor:

Upcoming Conference

ICSE 2025

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)7
  • Downloads (Last 6 weeks)1
Reflects downloads up to 06 Oct 2024

Other Metrics

Citations

Cited By

View all
  • (2022)MULA: A Just-In-Time Multi-labeling System for Issue ReportsIEEE Transactions on Reliability10.1109/TR.2021.307451271:1(250-263)Online publication date: Mar-2022
  • (2022)Detecting non-natural language artifacts for de-noising bug reportsAutomated Software Engineering10.1007/s10515-022-00350-029:2Online publication date: 1-Nov-2022
  • (2021)Just-in-time defect prediction for Android apps via imbalanced deep learning modelProceedings of the 36th Annual ACM Symposium on Applied Computing10.1145/3412841.3442019(1447-1454)Online publication date: 22-Mar-2021
  • (2021)Identifying non-natural language artifacts in bug reports2021 36th IEEE/ACM International Conference on Automated Software Engineering Workshops (ASEW)10.1109/ASEW52652.2021.00046(191-197)Online publication date: Nov-2021
  • (2020)Root cause prediction based on bug reports2020 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)10.1109/ISSREW51248.2020.00067(171-176)Online publication date: Oct-2020
  • (2020)A Fault Localization and Debugging Support Framework driven by Bug Tracking Data2020 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)10.1109/ISSREW51248.2020.00053(139-142)Online publication date: Oct-2020
  • (2020)Prediction Priority of Defective Modules for Testing Resource AllocationAutomated Software Testing10.1007/978-981-15-2455-4_5(95-109)Online publication date: 4-Feb-2020
  • (2018)Smart contracts vulnerabilities: a call for blockchain software engineering?2018 International Workshop on Blockchain Oriented Software Engineering (IWBOSE)10.1109/IWBOSE.2018.8327567(19-25)Online publication date: 20-Mar-2018
  • (2017)CitySenseProceedings of the XP2017 Scientific Workshops10.1145/3120459.3120472(1-5)Online publication date: 22-May-2017

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media