Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

An uncertainty based incremental learning for identifying the severity of bug report

  • Original Article
  • Published:
International Journal of Machine Learning and Cybernetics Aims and scope Submit manuscript

Abstract

To ensure the reliability of software system, software developers have to keep track of the severity of bug reports, and fix critical bugs as soon as possible. Recently, automatic methods to identify the severity of bug reports have emerged as a promising tool to lessen the work burden of software developers. However, most of such methods are supervised and data-driven models which fail to provide favorable performance in the presence of insufficient labeled sample or limited training data. In order to tackle with these issues, we propose an incremental learning for bug reports recognition. According to this framework of incremental learning, one active learning method is developed for tagging unlabeled bug reports, meanwhile, a sample augmentation method is utilized for sufficient training data. Both of these methods are based on uncertainty which is correlated to the informativeness and the classification risk of samples. Moreover, different types of connectionist models are employed to identify bug reports, and comprehensive experiments on real bug report datasets demonstrate that the generalization abilities of these models can be improved by this proposed incremental learning.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Hinton G, Deng L, Yu D, Dahl GE, Mohamed AR, Jaitly N, Kingsbury B (2012) Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process Mag 29(6):82–97

    Article  Google Scholar 

  2. Johnson J, Alahi A, L FF (2016) Perceptual losses for real-time style transfer and super-resolution. In: Proceedings of the European Conference on Computer Vision, pp 694–711

    Chapter  Google Scholar 

  3. Cho K, Van MB, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv:1406.1078

  4. Han K, Yu D, Tashev I (2014) Speech emotion recognition using deep neural network and extreme learning machine. In: Proceedings of the fifteenth annual conference of the international speech communication association, pp 223–227

  5. Zeng Y, Xu X, Fang Y, Zhao K (2015) Traffic sign recognition using extreme learning classifier with deep convolutional features. In: Proceedings of the international conference on intelligence science and big data engineering, pp 272–280

  6. Huang GB, Zhou H, Ding X, Zhang R (2012) Extreme learning machine for regression and multiclass classification. IEEE Trans Syst, Man, Cybern, Part B (Cybernetics) 42(2):513–529

    Article  Google Scholar 

  7. Wang C, Wang JH, Gu S (2016) Deep network based on stacked orthogonal convex incremental ELM autoencoders. Mathematical Problems in Engineering. https://doi.org/10.1155/2016/1649486

    MathSciNet  MATH  Google Scholar 

  8. Tang J, Deng C, Huang GB (2016) Extreme learning machine for multilayer perceptron. IEEE Trans Neural Netw Learn Syst 27(4):809–821

    Article  MathSciNet  Google Scholar 

  9. Huang GB, Bai Z, Kasun LLC, Vong CM (2015) Local receptive fields based extreme learning machine. IEEE Comput Intell Mag 10(2):18–29

    Article  Google Scholar 

  10. Cui C, Wang D (2016) High dimensional data regression using Lasso model and neural networks with random weights. Inf Sci 372:505–517

    Article  Google Scholar 

  11. Wang XZ, Zhang T, Wang R (2017) Noniterative deep learning: incorporating restricted boltzmann machine into multilayer random weight neural networks. IEEE Trans Syst, Man, Cybern: Syst 99:1–10

    Google Scholar 

  12. Chen CP, Liu Z (2018) Broad learning system: an effective and efficient incremental learning system without the need for deep architecture. IEEE Trans Neural Netw Learn Syst 29(1):10–24

    Article  MathSciNet  Google Scholar 

  13. Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems pp 1097–1105

  14. Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556

  15. Deng J, Dong W, Socher R, Li LJ, Li K, L FF (2009) Imagenet: A large-scale hierarchical image database. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp 248–255

  16. Thomee B, Shamma DA, Friedland G, Elizalde B, Ni K, Poland, D (2015) The new data and new challenges in multimedia research. arXiv:1503.01817

  17. Mao K, Capra L, Harman M, Jia Y (2015) A survey of the use of crowdsourcing in software engineering. J Syst Softw 126:57–84

    Article  Google Scholar 

  18. Menzies T, Marcus A (2008) Automated severity assessment of software defect reports. In: Proceedings of IEEE International Conference on Software Maintenance, pp 346–355

  19. Wang R, Chen DG, Kwong S (2014) Fuzzy rough set based active learning. IEEE Trans Fuzzy Syst 22(6):1699–1704

    Article  Google Scholar 

  20. Wang R, Wang XZ, Kwong S, Xu C (2017) Incorporating diversity and informativeness in multiple-instance active learning. IEEE Trans Fuzzy Syst 25(6):1460–1475

    Article  Google Scholar 

  21. Zhang T, Chen J, Yang G, Lee B, Luo X (2016) Towards more accurate severity prediction and fixer recommendation of software bugs. J Syst Softw 117:166–184

    Article  Google Scholar 

  22. Lamkanfi A, Demeyer S, Giger E, Goethals B (2010) Predicting the severity of a reported bug. In: Proceedings of the 7th IEEE Working Conference on Mining Software Repositories, pp 1–10

  23. Lamkanfi A, Demeyer S, Soetens QD, Verdonck T (2011) Comparing mining algorithms for predicting the severity of a reported bug. In: Proceedings of the 15th European Conference on Software Maintenance and Reengineering, pp 249–258

  24. Hinton GE, Srivastava N, Krizhevsky A, Sutskever I, Salakhutdinov RR (2012) Improving neural networks by preventing co-adaptation of feature detectors. arXiv:1207.0580

  25. Wang JJ, Lu Y (2014) Coordinate descent based hierarchical interactive lasso penalized logistic regression and its application to classification problems. Mathematical Problems in Engineering. https://doi.org/10.1155/2014/430201

    MATH  Google Scholar 

  26. Zhang Z, Luo P, Loy CC, Tang X (2014) Facial landmark detection by deep multi-task learning. In: Proceedings of the European Conference on Computer Vision, pp 94–108

    Chapter  Google Scholar 

  27. Pan SJ, Yang Q (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345–1359

    Article  Google Scholar 

  28. Hinton GE, Osindero S, Teh YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18(7):1527–1554

    Article  MathSciNet  Google Scholar 

  29. Wang R, Chow CY, Kwong S (2016) Ambiguity based multiclass active learning. IEEE Trans Fuzzy Syst 24(1):242–248

    Article  Google Scholar 

  30. Wang XZ, Wang R, Xu R (2018) Discovering the relationship between generalization and uncertainty by incorporating complexity of classification. IEEE Trans Cybern 48(2):703–715

    Article  MathSciNet  Google Scholar 

  31. Schez D, Trillas E (2012) Measures of fuzziness under different uses of fuzzy sets. In: Proceedings of International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems, pp 25–34

  32. Wang XZ, Xing HJ, Li Y, Hua Q, Dong CR, Pedrycz W (2015) A study on relationship between generalization abilities and fuzziness of base classifiers in ensemble learning. IEEE Trans Fuzzy Syst 23(5):1638–1654

    Article  Google Scholar 

  33. Zamir AR, Sax A, Shen W, Guibas L, Malik J, Savarese S (2018) Taskonomy: Disentangling Task Transfer Learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 3712–3722

  34. Eclipse, http://bugs.eclipse.org/bugs, 2/2/2018 available

  35. Mozilla, http://bugzilla.mozilla.org, 2/2/2018 available

  36. GNOME, http://bugzilla.gnome.org, 2/2/2018 available

  37. Bugzilla, https://www.bugzilla.org/, 2/2/2018 available

  38. Zhu HY, Wang XZ (2017) A cost-sensitive semi-supervised learning model based on uncertainty. Neurocomputing 251:106–114

    Article  Google Scholar 

  39. Hinton GE (2002) Training products of experts by minimizing contrastive divergence. Neural Comput 14(8):1771–1800

    Article  Google Scholar 

  40. Tieleman T, Hinton G (2009). Using fast weights to improve persistent contrastive divergence. In: Proceedings of the 26th Annual International Conference on Machine Learning, pp 1033–1040

  41. Salakhutdinov R, Larochelle H (2010) Efficient learning of deep Boltzmann machines. In: Proceedings of the 13th international conference on artificial intelligence and statistics, pp 693–700

  42. Kirkpatrick J, Pascanu R, Rabinowitz N, Veness J, Desjardins G, Rusu AA (2017) Overcoming catastrophic forgetting in neural networks. In: Proceedings of the national academy of sciences, pp 3521–3526

    Article  MathSciNet  Google Scholar 

  43. Dai WY, Yang Q, Xue GR, Yu Y (2007) Boosting for transfer learning. In: Proceedings of the 24th international conference on Machine learning, pp 193–200

  44. Zhai JH, Zhang SF, Wang CX (2017) The classification of imbalanced large data sets based on mapreduce and ensemble of elm classifiers. Int J Mach Learn Cybern 8(3):1009–1017

    Article  Google Scholar 

  45. Zhai JH, Wang XZ, Pang XH (2016) Voting-based instance selection from large data sets with MapReduce and random weight networks. Inf Sci 367:1066–1077

    Article  Google Scholar 

  46. Zhai JH, Li T, Wang XZ (2016) A cross-selection instance algorithm. J Intell Fuzzy Syst 30(2):717–728

    Article  Google Scholar 

  47. Wang XZ, Ashfaq RAR, Fu AM (2015) Fuzziness based sample categorization for classifier performance improvement. J Intell Fuzzy Syst 29(3):1–12

    MathSciNet  Google Scholar 

  48. De Luca A, Termini S (1993) A definition of a nonprobabilistic entropy in the setting of fuzzy sets theory. In Readings in Fuzzy Sets for Intelligent Systems 197–202

  49. Schez D, Trillas E (2012) Measures of fuzziness under different uses of fuzzy sets. In: Proceedings of the International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems, pp 25–34

  50. De Luca A, Termini S (1974) Entropy of L-fuzzy sets. Inf Control 24(1):55–73

    Article  MathSciNet  Google Scholar 

  51. Zadeh LA (1968) Probability measures of fuzzy events. J Math Anal Appl 23(2):421–427

    Article  MathSciNet  Google Scholar 

  52. Bengio Y, Delalleau O (2009) Justifying and generalizing contrastive divergence. Neural Comput 21(6):1601–1621

    Article  MathSciNet  Google Scholar 

  53. Bengio Y, Lamblin P, Popovici D, Larochelle H (2007) Greedy layer-wise training of deep networks. In Advances in neural information processing systems 153–160

  54. Hinton GE (2007) Learning multiple layers of representation. Trends Cognit Sci 11(10):428–434

    Article  Google Scholar 

Download references

Acknowledgements

This work is supported by the National Natural Science Foundation of China under Grant 61672122, Grant 61602077, the Public Welfare Funds for Scientific Research of Liaoning Province of China under Grant 20170005, the Natural Science Foundation of Liaoning Province of China under Grant 20170540097, and the Fundamental Research Funds for the Central Universities under Grant 3132016348.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xi Yang.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

Given a RBM, we set the hidden vector without \(h_k\) as \(\mathbf{h }_{\tilde{k}}=(h_1,h_2, ..., h_{k-1}, h_{k+1}, ..., h_{n_h})^T\), where \(h_i\) is the ith hidden node. When the \(h_k\) is 1, the energy function (9) can be split into two items as follows (Fig. 7):

$$\begin{aligned} \alpha (\mathbf{v })= & {} b_k+\sum _{i=1}^{n_v}w_{k, i}v_i \end{aligned}$$
(23)
$$\begin{aligned} \beta (\mathbf{v }, \mathbf{h }_{\tilde{k}})= & {} \sum _{i=1}^{n_v}a_iv_i+\sum _{j=1,j \ne k}^{n_h}b_jh_j+\sum _{i=1}^{n_v}\sum _{j=1, j\ne k}^{n_h}h_jw_{j,i}v_i. \end{aligned}$$
(24)

Then, the derivation process of conditional probability \(P(h_k=1|\mathbf{v })\) is shown below

$$\begin{aligned} \begin{aligned} p(h_k=1|\mathbf{v })&= p(h_k=1|\mathbf{h }_{\tilde{k}}, \mathbf{v }) \\&= \frac{p(h_k=1, \mathbf{h }_{\tilde{k}}, \mathbf{v })}{p(\mathbf{h }_{\tilde{k}}, \mathbf{v })} \\&= \frac{p(h_k=1, \mathbf{h }_{\tilde{k}}, \mathbf{v })}{p(h_k=1, \mathbf{h }_{\tilde{k}}, \mathbf{v })+p(h_k=0, \mathbf{h }_{\tilde{k}}, \mathbf{v })} \\&= \frac{\frac{1}{Z}e^{-E(h_k=1, \mathbf{h }_{\tilde{k}}, \mathbf{v })}}{\frac{1}{Z}e^{-E(h_k=1, \mathbf{h }_{\tilde{k}}, \mathbf{v })}+\frac{1}{Z}e^{-E(h_k=0, \mathbf{h }_{\tilde{k}}, \mathbf{v })}} \\&= \frac{e^{-E(h_k=1, \mathbf{h }_{\tilde{k}}, \mathbf{v })}}{e^{-E(h_k=1, \mathbf{h }_{\tilde{k}}, \mathbf{v })}+e^{-E(h_k=0, \mathbf{h }_{\tilde{k}}, \mathbf{v })}}\\&= \frac{1}{1+e^{-E(h_k=0,\mathbf{h }_{\tilde{k}},\mathbf{v })+E(h_k=1,\mathbf{h }_{\tilde{k}},\mathbf{v })}}\\&= \frac{1}{1+e^{[\beta (\mathbf{h }_{\tilde{k}},\mathbf{v }+0 \cdot \alpha _{k}(\mathbf{v }))]+[-\beta (\mathbf{h }_{\tilde{k}},\mathbf{v })+1 \cdot \alpha _{k}(\mathbf{v })]}}\\&= \frac{1}{1+e^{-\alpha _{k}(\mathbf{v })}} = sigmoid(\alpha _{k}(\mathbf{v })) \end{aligned} \end{aligned}$$
(25)

The derivation process of \(p(v_k=1|\mathbf{h })\) is the same way as shown above.

Fig. 7
figure 7figure 7figure 7

The overall results of incremental learning. The blue lines are the results of CELM, and the green lines are the results of BLS

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, TL., Chen, R., Yang, X. et al. An uncertainty based incremental learning for identifying the severity of bug report. Int. J. Mach. Learn. & Cyber. 11, 123–136 (2020). https://doi.org/10.1007/s13042-019-00961-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13042-019-00961-2

Keywords