Abstract
To ensure the reliability of software system, software developers have to keep track of the severity of bug reports, and fix critical bugs as soon as possible. Recently, automatic methods to identify the severity of bug reports have emerged as a promising tool to lessen the work burden of software developers. However, most of such methods are supervised and data-driven models which fail to provide favorable performance in the presence of insufficient labeled sample or limited training data. In order to tackle with these issues, we propose an incremental learning for bug reports recognition. According to this framework of incremental learning, one active learning method is developed for tagging unlabeled bug reports, meanwhile, a sample augmentation method is utilized for sufficient training data. Both of these methods are based on uncertainty which is correlated to the informativeness and the classification risk of samples. Moreover, different types of connectionist models are employed to identify bug reports, and comprehensive experiments on real bug report datasets demonstrate that the generalization abilities of these models can be improved by this proposed incremental learning.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Hinton G, Deng L, Yu D, Dahl GE, Mohamed AR, Jaitly N, Kingsbury B (2012) Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process Mag 29(6):82–97
Johnson J, Alahi A, L FF (2016) Perceptual losses for real-time style transfer and super-resolution. In: Proceedings of the European Conference on Computer Vision, pp 694–711
Cho K, Van MB, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv:1406.1078
Han K, Yu D, Tashev I (2014) Speech emotion recognition using deep neural network and extreme learning machine. In: Proceedings of the fifteenth annual conference of the international speech communication association, pp 223–227
Zeng Y, Xu X, Fang Y, Zhao K (2015) Traffic sign recognition using extreme learning classifier with deep convolutional features. In: Proceedings of the international conference on intelligence science and big data engineering, pp 272–280
Huang GB, Zhou H, Ding X, Zhang R (2012) Extreme learning machine for regression and multiclass classification. IEEE Trans Syst, Man, Cybern, Part B (Cybernetics) 42(2):513–529
Wang C, Wang JH, Gu S (2016) Deep network based on stacked orthogonal convex incremental ELM autoencoders. Mathematical Problems in Engineering. https://doi.org/10.1155/2016/1649486
Tang J, Deng C, Huang GB (2016) Extreme learning machine for multilayer perceptron. IEEE Trans Neural Netw Learn Syst 27(4):809–821
Huang GB, Bai Z, Kasun LLC, Vong CM (2015) Local receptive fields based extreme learning machine. IEEE Comput Intell Mag 10(2):18–29
Cui C, Wang D (2016) High dimensional data regression using Lasso model and neural networks with random weights. Inf Sci 372:505–517
Wang XZ, Zhang T, Wang R (2017) Noniterative deep learning: incorporating restricted boltzmann machine into multilayer random weight neural networks. IEEE Trans Syst, Man, Cybern: Syst 99:1–10
Chen CP, Liu Z (2018) Broad learning system: an effective and efficient incremental learning system without the need for deep architecture. IEEE Trans Neural Netw Learn Syst 29(1):10–24
Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems pp 1097–1105
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556
Deng J, Dong W, Socher R, Li LJ, Li K, L FF (2009) Imagenet: A large-scale hierarchical image database. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp 248–255
Thomee B, Shamma DA, Friedland G, Elizalde B, Ni K, Poland, D (2015) The new data and new challenges in multimedia research. arXiv:1503.01817
Mao K, Capra L, Harman M, Jia Y (2015) A survey of the use of crowdsourcing in software engineering. J Syst Softw 126:57–84
Menzies T, Marcus A (2008) Automated severity assessment of software defect reports. In: Proceedings of IEEE International Conference on Software Maintenance, pp 346–355
Wang R, Chen DG, Kwong S (2014) Fuzzy rough set based active learning. IEEE Trans Fuzzy Syst 22(6):1699–1704
Wang R, Wang XZ, Kwong S, Xu C (2017) Incorporating diversity and informativeness in multiple-instance active learning. IEEE Trans Fuzzy Syst 25(6):1460–1475
Zhang T, Chen J, Yang G, Lee B, Luo X (2016) Towards more accurate severity prediction and fixer recommendation of software bugs. J Syst Softw 117:166–184
Lamkanfi A, Demeyer S, Giger E, Goethals B (2010) Predicting the severity of a reported bug. In: Proceedings of the 7th IEEE Working Conference on Mining Software Repositories, pp 1–10
Lamkanfi A, Demeyer S, Soetens QD, Verdonck T (2011) Comparing mining algorithms for predicting the severity of a reported bug. In: Proceedings of the 15th European Conference on Software Maintenance and Reengineering, pp 249–258
Hinton GE, Srivastava N, Krizhevsky A, Sutskever I, Salakhutdinov RR (2012) Improving neural networks by preventing co-adaptation of feature detectors. arXiv:1207.0580
Wang JJ, Lu Y (2014) Coordinate descent based hierarchical interactive lasso penalized logistic regression and its application to classification problems. Mathematical Problems in Engineering. https://doi.org/10.1155/2014/430201
Zhang Z, Luo P, Loy CC, Tang X (2014) Facial landmark detection by deep multi-task learning. In: Proceedings of the European Conference on Computer Vision, pp 94–108
Pan SJ, Yang Q (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345–1359
Hinton GE, Osindero S, Teh YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18(7):1527–1554
Wang R, Chow CY, Kwong S (2016) Ambiguity based multiclass active learning. IEEE Trans Fuzzy Syst 24(1):242–248
Wang XZ, Wang R, Xu R (2018) Discovering the relationship between generalization and uncertainty by incorporating complexity of classification. IEEE Trans Cybern 48(2):703–715
Schez D, Trillas E (2012) Measures of fuzziness under different uses of fuzzy sets. In: Proceedings of International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems, pp 25–34
Wang XZ, Xing HJ, Li Y, Hua Q, Dong CR, Pedrycz W (2015) A study on relationship between generalization abilities and fuzziness of base classifiers in ensemble learning. IEEE Trans Fuzzy Syst 23(5):1638–1654
Zamir AR, Sax A, Shen W, Guibas L, Malik J, Savarese S (2018) Taskonomy: Disentangling Task Transfer Learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 3712–3722
Eclipse, http://bugs.eclipse.org/bugs, 2/2/2018 available
Mozilla, http://bugzilla.mozilla.org, 2/2/2018 available
GNOME, http://bugzilla.gnome.org, 2/2/2018 available
Bugzilla, https://www.bugzilla.org/, 2/2/2018 available
Zhu HY, Wang XZ (2017) A cost-sensitive semi-supervised learning model based on uncertainty. Neurocomputing 251:106–114
Hinton GE (2002) Training products of experts by minimizing contrastive divergence. Neural Comput 14(8):1771–1800
Tieleman T, Hinton G (2009). Using fast weights to improve persistent contrastive divergence. In: Proceedings of the 26th Annual International Conference on Machine Learning, pp 1033–1040
Salakhutdinov R, Larochelle H (2010) Efficient learning of deep Boltzmann machines. In: Proceedings of the 13th international conference on artificial intelligence and statistics, pp 693–700
Kirkpatrick J, Pascanu R, Rabinowitz N, Veness J, Desjardins G, Rusu AA (2017) Overcoming catastrophic forgetting in neural networks. In: Proceedings of the national academy of sciences, pp 3521–3526
Dai WY, Yang Q, Xue GR, Yu Y (2007) Boosting for transfer learning. In: Proceedings of the 24th international conference on Machine learning, pp 193–200
Zhai JH, Zhang SF, Wang CX (2017) The classification of imbalanced large data sets based on mapreduce and ensemble of elm classifiers. Int J Mach Learn Cybern 8(3):1009–1017
Zhai JH, Wang XZ, Pang XH (2016) Voting-based instance selection from large data sets with MapReduce and random weight networks. Inf Sci 367:1066–1077
Zhai JH, Li T, Wang XZ (2016) A cross-selection instance algorithm. J Intell Fuzzy Syst 30(2):717–728
Wang XZ, Ashfaq RAR, Fu AM (2015) Fuzziness based sample categorization for classifier performance improvement. J Intell Fuzzy Syst 29(3):1–12
De Luca A, Termini S (1993) A definition of a nonprobabilistic entropy in the setting of fuzzy sets theory. In Readings in Fuzzy Sets for Intelligent Systems 197–202
Schez D, Trillas E (2012) Measures of fuzziness under different uses of fuzzy sets. In: Proceedings of the International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems, pp 25–34
De Luca A, Termini S (1974) Entropy of L-fuzzy sets. Inf Control 24(1):55–73
Zadeh LA (1968) Probability measures of fuzzy events. J Math Anal Appl 23(2):421–427
Bengio Y, Delalleau O (2009) Justifying and generalizing contrastive divergence. Neural Comput 21(6):1601–1621
Bengio Y, Lamblin P, Popovici D, Larochelle H (2007) Greedy layer-wise training of deep networks. In Advances in neural information processing systems 153–160
Hinton GE (2007) Learning multiple layers of representation. Trends Cognit Sci 11(10):428–434
Acknowledgements
This work is supported by the National Natural Science Foundation of China under Grant 61672122, Grant 61602077, the Public Welfare Funds for Scientific Research of Liaoning Province of China under Grant 20170005, the Natural Science Foundation of Liaoning Province of China under Grant 20170540097, and the Fundamental Research Funds for the Central Universities under Grant 3132016348.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
Given a RBM, we set the hidden vector without \(h_k\) as \(\mathbf{h }_{\tilde{k}}=(h_1,h_2, ..., h_{k-1}, h_{k+1}, ..., h_{n_h})^T\), where \(h_i\) is the ith hidden node. When the \(h_k\) is 1, the energy function (9) can be split into two items as follows (Fig. 7):
Then, the derivation process of conditional probability \(P(h_k=1|\mathbf{v })\) is shown below
The derivation process of \(p(v_k=1|\mathbf{h })\) is the same way as shown above.
Rights and permissions
About this article
Cite this article
Zhang, TL., Chen, R., Yang, X. et al. An uncertainty based incremental learning for identifying the severity of bug report. Int. J. Mach. Learn. & Cyber. 11, 123–136 (2020). https://doi.org/10.1007/s13042-019-00961-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13042-019-00961-2