Abstract
Prediction occurs in many biological nervous systems e.g. in the cortex [7]. We introduce a method of adapting the recurrent layer dynamics of an echo-state network (ESN) without attempting to train the weights directly. Initially a network is generated that fulfils the echo state – liquid state condition. A second network is then trained to predict the next internal state of the system. In simulation, the prediction of this module is then mixed with the actual activation of the internal state neurons, to produce dynamics that a partially driven by the network model, rather than the input data. The mixture is determined by a parameter α. The target function be produced by the network was sin 3(0.24t), given an input function sin(0.24t). White noise was added to the input signal at 15% of the amplitude of the signal. Preliminary results indicate that self prediction may improve performance of an ESN when performing signal mappings in the presence of additive noise.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Jaeger, H.: The ’echo state’ approach to analysing and training recurrent neural networks. GMD Report 148, GMD German National Research Institute for Computer Science (2001), http://www.gmd.de/People/Herbert.Jaeger/Publications.html
Jaeger, H.: Short term memory in echo state networks. GMD Report 152, GMD German National Research Insitute for Computer Science (2001), http://www.gmd.de/People/Herbert.Jaeger/Publications.html
Jaeger, H.: Adaptive nonlinear system identification with echo state networks. In: Proc. of NIPS 2002, AA14 (2003)
Maas, W., Natschläger, T., Markram, H.: Real-time computing without stable states: A new framework on for neural computation based on perturbations. NeuroCOLT Technical Report NC-TR-01-113 (2001)
Farhang-Boroujeny, B.: Adaptive Filters. Wiley, Chichester (1999)
Werbos, P.: Backpropagation through time: what it does and how to do it. Proceedings of the IEEE 78(10), 1550–1560 (1990)
Huettel, S., Mack, P., McCarthy, G.: Perceiving patterns in random series: dynamic processing of sequence in the prefrontal cortex. Nature Neurosci 5(5), 485–490 (2002)
Gers, F.A., Schmidhuber, J., Cummins, F.: Learning to forget: continual prediction with LSTM. Neural Computation 12(10), 2451–2471 (2000)
Elman, J.: Finding Structure in Time. Cognitive Science 14(2), 179–211 (1990)
Bengio, Y.: Neural Networks for Speech and Sequence Recognition. International Thomson Publishing Inc. (1996)
Kadous, W.: Temporal Classification: Extending the Classification Paradigm to Multivariate Time Series Dissertation, School of Computer Science and Engineering, University of New South Wales (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Mayer, N.M., Browne, M. (2004). Echo State Networks and Self-Prediction. In: Ijspeert, A.J., Murata, M., Wakamiya, N. (eds) Biologically Inspired Approaches to Advanced Information Technology. BioADIT 2004. Lecture Notes in Computer Science, vol 3141. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-27835-1_4
Download citation
DOI: https://doi.org/10.1007/978-3-540-27835-1_4
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-23339-8
Online ISBN: 978-3-540-27835-1
eBook Packages: Springer Book Archive