Abstract
The task rehearsal method (TRM) is introduced as an approach to life-long learning that uses the representation of previously learned tasks as a source of inductive bias. This inductive bias enables TRM to generate more accurate hypotheses for new tasks that have small sets of training examples. TRM has a knowledge retention phase during which the neural network representation of a successfully learned task is stored in a domain knowledge database, and a knowledge recall and learning phase during which virtual examples of stored tasks are generated from the domain knowledge. The virtual examples are rehearsed as secondary tasks in parallel with the learning of a new (primary) task using the ŋMTL neural network algorithm, a variant of multiple task learning (MTL). The results of experiments on three domains show that TRM is effective in retaining task knowledge in a representational form and transferring that knowledge in the form of virtual examples. TRM with ŋMTL is shown to develop more accurate hypotheses for tasks that suffer from impoverished training sets.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Yaser S. Abu-Mostafa. Hints. Neural Computation, 7:639–671, 1995.
Jonathan Baxter. Learning internal representations. Proceedings of the Eighth International Conference on Computational Learning Theory, 1995.
Richard A. Caruana. Multitask learning. Machine Learning, 28:41–75, 1997.
Michael McCloskey and Neal J. Cohen. Catastrophic interference in connectionist networks: the sequential learning problem. The Psychology of Learning and Motivation, 24:109–165, 1989.
Tom M. Mitchell. Machine Learning. McGraw Hill, New York, NY, 1997.
Anthony V. Robins. Catastrophic forgetting, rehearsal, and pseudorehearsal. Connection Science, 7:123–146, 1995.
Daniel L. Silver. Selective Transfer of Neural Network Task Knowledge. PhD Thesis, Dept. of Computer Science, University of Western Ontario, London, Canada, June 2000.
Daniel L. Silver and Robert E. Mercer. The parallel transfer of task knowledge using dynamic learning rates based on a measure of relatedness. Connection Science Special Issue: Transfer in Inductive Systems, 8(2):277–294, 1996.
Daniel L. Silver and Robert E. Mercer. Selective functional transfer: Inductive bias from related tasks. Proceedings of the International Conference on Artificial Intelligence and Soft Computing (ASC2001), pages 182–189, 2001.
Sebastian Thrun. Lifelong learning algorithms. Learning to Learn, pages 181–209, 1997.
Paul E. Utgoff. Machine Learning of Inductive Bias. Kluwer Academc Publisher, Boston, MA, 1986.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Silver, D.L., Mercer, R.E. (2002). The Task Rehearsal Method of Life-Long Learning: Overcoming Impoverished Data. In: Cohen, R., Spencer, B. (eds) Advances in Artificial Intelligence. Canadian AI 2002. Lecture Notes in Computer Science(), vol 2338. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-47922-8_8
Download citation
DOI: https://doi.org/10.1007/3-540-47922-8_8
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-43724-6
Online ISBN: 978-3-540-47922-2
eBook Packages: Springer Book Archive