Abstract
Most of the existing SNNs only consider training the noise-free data. However, noise extensively exists in actual SNNs. The stability of networks is affected by noise perturbation during the training period. Therefore, one research challenge is to improve the stability and produce reliable outputs under the present of noises. In this paper, the training method and the exponential method are employed to enhance the neural network ability of noise tolerance. The comparison of conventional and anti-noise SNNs under various tasks shows that the anti-noise SNN can significantly improve the noise tolerance capability.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Charlotte, Y.-F.H., Bingo, W.-K.L.: Design of multi-layer perceptrons via joint filled function and genetic algorithm approach for video forensics. In: IEEE International Conference on Consumer Electronics-China (ICCE-China), pp. 1–4 (2016)
Fei, Y., Hu, J., Gao, K., Tu, J., Li, W., Wang, W.: Predicting risk for portal vein thrombosis in acute pancreatitis patients: A comparison of radical basis function artificial neural network and logistic regression models. J. Crit. Care 39, 115–123 (2017)
Wang, J., Wen, Y., Ye, Z., Jian, L., Chen, H.: Convergence analysis of BP neural networks via sparse response. Appl. Soft. Comput. 61, 354–363 (2017)
Fumihiro, K., Kato, S., Nakamura, M.: Multi-task reinforcement learning with associative memory models considering the multiple distributions of MDPs. In: Global Conference on Consumer Electronics (GCCE), pp. 27–29 (2015)
Blessen, C.E., Douglas, P.M., David, X.C.: Neuroprosthetics in amputee and brain injury rehabilitation. Exp. Neurol. 287, 479–485 (2017)
James, C.D., et al.: A historical survey of algorithms and hardware architectures for neural-inspired and neuromorphic computing applications. Biol. Inspir. Cognit. Archit. 19, 49–64 (2017)
Liu, J., Harkin, J., Maguire, L.P., McDaid, L., Wade, J.: SPANNER: A self-repairing spiking neural network hardware architecture. In: IEEE Transactions on Neural Networks and Learning Systems, pp. 1–14 (2017)
Liu, J., Harkin, J., Mcelholm, M., McDaid, L.: Case study : Bio-inspired self-adaptive strategy for spike-based PID controller. In: IEEE International Symposium on Circuits and Systems, pp. 2700–2703 (2015)
Liu, J., Harkin, J., Mcdaid, L., Halliday, D.M., Tyrrell, A.M., Timmis, J.: Self-repairing mobile robotic car using astrocyte-neuron networks. In: International Joint Conference on Neural Networks, pp. 1379–1386 (2016)
Shrestha, S.B., Song, Q.: Robust spike-train learning in spike-event based weight update. Neural Netw 96, 33–46 (2017)
Yang, J., Yang, W., Wu, W.: A remark on the error-backpropagation learning algorithm for spiking. Appl. Math. Lett. 25(8), 1118–1120 (2012)
Bao, H., Cao, J.: Delay-distribution-dependent state estimation for discrete-time stochastic neural networks with random delay. Neural Netw. 24(1), 19–28 (2011)
Zhang, M., Li, J., Wang, Y., Gao, G.: R-tempotron: A robust tempotron learning rule for spike timing-based decisions. In: International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), pp. 139–142 (2016)
Kawanishi, K., Takase, H., Kawanaka, H., Tsuruoka, S.: Reduce the computing time for SpikeProp by approximation of spike response function. In: 20th International Conference on Knowledge Based and Intelligent Information and Engineering Systems, vol. 96, pp. 1186–1192 (2016)
Booij, O., Tat Nguyen, H.: A gradient descent rule for spiking neurons emitting multiple spikes. Inf. Process. Lett. 95(6), 552–558 (2005)
Bohte, S.M., Kok, J.N., La Poutré, H.: Error-backpropagation in temporally encoded networks of spiking neurons. Neurocomputing 48(1–4), 17–37 (2002)
Luo, Y., Fu, Q., Liu, J., Harkin, J.: An extended algorithm using adaptation of momentum and learning rate for spiking neurons emitting multiple spikes. Int. Work-Conf. Artif. Neural Netw. (IWANN) 237, 569–579 (2017)
Fu, Q., et al.: Improving learning algorithm performance for spiking neural networks. In: IEEE 17th International Conference on Communication Technology (ICCT), pp. 1916–1919 (2017)
Schrauwen, B., Van Campenhout, J.: Extending spikeprop. In: International Joint Conference on Neural Networks, pp. 471–476 (2004)
Acknowledgement
This research was supported by the National Natural Science Foundation of China under Grant 61603104, the Guangxi Natural Science Foundation under Grants 2016GXNSFCA380017, 2015GXNSFBA139256 and 2017GXNSFAA198180, the funding of Overseas 100 Talents Program of Guangxi Higher Education, and the Doctoral Research Foundation of Guangxi Normal University under Grant 2016BQ005.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Luo, Y., Fu, Q., Liu, J., Huang, Y., Ding, X., Cao, Y. (2018). Improving the Stability for Spiking Neural Networks Using Anti-noise Learning Rule. In: Geng, X., Kang, BH. (eds) PRICAI 2018: Trends in Artificial Intelligence. PRICAI 2018. Lecture Notes in Computer Science(), vol 11013. Springer, Cham. https://doi.org/10.1007/978-3-319-97310-4_4
Download citation
DOI: https://doi.org/10.1007/978-3-319-97310-4_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-97309-8
Online ISBN: 978-3-319-97310-4
eBook Packages: Computer ScienceComputer Science (R0)