Abstract
This research work presents a methodology for debugging embedded designs by using recurrent neural networks. In this methodology, a cycle-accurate lossless debugging system with unlimited trace window is used for debugging. The lossless trace resembles a time data series. A recurrent neural network trained either through a golden reference or from the actual time series can be used to predict the incoming debugging data. A bug can be easily isolated based upon the discrepancy between the received and the predicted time series. This allows to draw conclusions to speed up the debugging process.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Huang, C., Yin, Y., Hsu, C., Huang, T.B., Chang, T.: SoC HW/SW verification and validation. In: 16th Asia and South Pacific Design Automation Conference (ASP-DAC 2011), pp. 297–300, January 2011
Fania, M., Peiravi, P., Chandramouly, A., Yalla, C.: White paper: predictive analytics and interactive queries on big data. Intel (2013)
Jindal, A., Kumar, B., Jindal, N., Fujita, M., Singh, V.: Silicon debug with maximally expanded internal observability using nearest neighbor algorithm. In: 2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI), pp. 46–51, July 2018
Wang, L.: Experience of data analytics in EDA and test principles, promises, and challenges. IEEE Trans. Comput. Aided Des. Integr. Circuits Syst. 36(6), 885–898 (2017)
Mandouh, E., Wassal, A.G.: Application of machine learning techniques in post-silicon debugging and bug localization. J. Electron. Test. 34(2), 163–181 (2018)
Groß, W., Lange, S., Bödecker, J., Blum, M.: Predicting time series with space-time convolutional and recurrent neural networks. In: ESANN (2017)
Khan, H.H., Göhringer, D.: FPGA debugging by a device start and stop approach. In: 2016 International Conference on ReConFigurable Computing and FPGAs (ReConFig), pp. 1–6, November 2016
Petropoulos, F., Makridakis, S., Assimakopoulos, V., Nikolopoulos, K.: ‘Horses for Courses’ in demand forecasting. Eur. J. Oper. Res. 237(1), 152–163 (2014)
Friedman, J.H., Rafsky, L.C.: Multivariate generalizations of the Wald-Wolfowitz and Smirnov two-sample tests. Ann. Stat. 7(4), 697–717 (1979)
Chollet, F.: Deep Learning with Python, 1st edn. Manning Publications Co., Greenwich (2017)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980
Bianchi, F.M., Maiorino, E., Kampffmeyer, M.C., Rizzi, A., Jenssen, R.: Properties and training in recurrent neural networks. In: Recurrent Neural Networks for Short-Term Load Forecasting. SCS, pp. 9–21. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-70338-1_2
Masters, D., Luschi, C.: Revisiting small batch training for deep neural networks. arXiv preprint arXiv:1804.07612
Bertacco, V., Bonkowski, W.: ItHELPS: iterative high-accuracy error localization in post-silicon. In: 33rd IEEE International Conference on Computer Design (ICCD), pp. 196–199, October 2015
Acknowledgements
This research has received funding under the grant number 100369691 by the German Federal State of Saxony.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Khan, H., Podlubne, A., Akgün, G., Göhringer, D. (2020). Cycle-Accurate Debugging of Embedded Designs Using Recurrent Neural Networks. In: Rincón, F., Barba, J., So, H., Diniz, P., Caba, J. (eds) Applied Reconfigurable Computing. Architectures, Tools, and Applications. ARC 2020. Lecture Notes in Computer Science(), vol 12083. Springer, Cham. https://doi.org/10.1007/978-3-030-44534-8_6
Download citation
DOI: https://doi.org/10.1007/978-3-030-44534-8_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-44533-1
Online ISBN: 978-3-030-44534-8
eBook Packages: Computer ScienceComputer Science (R0)