Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Continual Learning by Task-Wise Shared Hidden Representation Alignment

  • Conference paper
  • First Online:
Artificial Neural Networks and Machine Learning – ICANN 2022 (ICANN 2022)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 13530))

Included in the following conference series:

  • 2575 Accesses

Abstract

One of the goals of machine learning is learning a sequence of tasks more naturally. Continual learning imitates the real learning mode of human beings and continually learns new knowledge without forgetting old knowledge at the same time. In the past decades, considerable attention has been paid to this learning method. However, avoiding forgetting old knowledge in the process of learning new knowledge remains an ongoing challenge due to catastrophic forgetting. Therefore, it is desirable to exploit a new method to improve the stability in a continual learning scenario. In this paper, we specifically focus on the shared feature extraction between two consecutive tasks. To this end, we explore a continual learning paradigm by using the task-wise shared hidden representation alignment module, which contrasts shared representations from the current task and shared representations from reconstruction pseudo samples of previous tasks. Our proposed TSHRA model grasps similarity features provided by the alignment module to learn shared representations more consummately when the model is learning the current task. To verify our proposed model, we conduct experiments on Split-MNIST and Fashion-MNIST. The experimental results show that our proposed TSHRA’s performance is outstanding which justify that the alignment module has a positive effect on learning shared representations among different tasks for the continual learning scenario.

J. Liu—This work was supported by the Science Foundation of China

University of Petroleum, Beijing (No. 2462020YXZZ023).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Ring, M.: Continual learning in reinforcement environment, Ph. D. thesis, University of Texas, Austin (1994)

    Google Scholar 

  2. French, R.M.: Catastrophic forgetting in connectionist networks. Trends Cogn. Sci. 3(4), 128–135 (1999)

    Google Scholar 

  3. Shim, D., et al.: Online class-incremental continual learning with adversarial shapley value. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35. no. 11 (2021)

    Google Scholar 

  4. Li, Z., Hoiem, D.: Learning without forgetting. IEEE Trans. Pattern Anal. Mach. Intell. 40(12), 2935–2947 (2017)

    Article  Google Scholar 

  5. Shin, H., et al.: Continual learning with deep generative replay. arXiv preprint arXiv:1705.08690 (2017)

  6. Kirkpatrick, J., et al.: Overcoming catastrophic forgetting in neural networks. Proc. Nat. Acad. Sci. 114(13), 3521–3526 (2017)

    Google Scholar 

  7. Fernando, C.: Evolution channels gradient descent in super neural networks. arXiv preprint arXiv:1701.08734 1–16 (2017)

  8. Zenke, F., Poole, B., Ganguli, S.: Continual learning through synaptic intelligence. In: International Conference on Machine Learning, PMLR (2017)

    Google Scholar 

  9. Van de Ven, G.M., Tolias, A.S.: Three scenarios for continual learning. arXiv preprint arXiv:1904.07734 (2019)

  10. Bromley, J., et al.: Signature verification using a “siamese” time delay neural network. Int. J. Pattern Recognit. Artif. Intell. 7(4), 669–688 (1993)

    Google Scholar 

  11. Ye, M., et al.: Unsupervised embedding learning via invariant and spreading instance feature. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2019)

    Google Scholar 

  12. Grill, J.-B., et al.: Bootstrap your own latent: a new approach to self-supervised learning. arXiv preprint arXiv:2006.07733 (2020)

  13. Chen, X., Kaiming, H.: Exploring simple siamese representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2021)

    Google Scholar 

  14. Chen, T., et al.: A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, PMLR (2020)

    Google Scholar 

  15. He, K., et al.: Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2020)

    Google Scholar 

  16. Wu, Z., et al.: Clear: contrastive learning for sentence representation. arXiv preprint arXiv:2012.15466 (2020)

  17. Wu, J., et al.: Self-supervised graph learning for recommendation. In: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (2021)

    Google Scholar 

  18. Kingma, D.P., Welling, M.: Auto-encoding variational Bayes. arXiv preprint arXiv:1312.6114 (2013)

  19. Goodfellow, I.J., et al.: Generative adversarial nets. In: Advances in Neural Information Processing systems 27 (2014)

    Google Scholar 

  20. Kingma, D.P., et al.: Semi-supervised learning with deep generative models. In: Advances in Neural Information Processing Systems (2014)

    Google Scholar 

  21. Rebuffi, S.-A., et al.: iCaRL: incremental classifier and representation learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)

    Google Scholar 

  22. Sokar, G., Mocanu, D.C., Pechenizkiy, M.: Learning invariant representation for continual learning (2021)

    Google Scholar 

  23. LeCun, Y.: The MNIST database of handwritten digits. http://yann.lecun.com/exdb/mnist/ (1998)

  24. Xiao, H., Rasul, K., Vollgraf, R.: Fashion-MNIST: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747 (2017)

  25. Lopez-Paz, D., Ranzato, M.: Gradient episodic memory for continual learning. Adv. Neural. Inf. Process. Syst. 30, 6467–6476 (2017)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jian-wei Liu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhan, Xh., Liu, Jw., Han, Yn. (2022). Continual Learning by Task-Wise Shared Hidden Representation Alignment. In: Pimenidis, E., Angelov, P., Jayne, C., Papaleonidas, A., Aydin, M. (eds) Artificial Neural Networks and Machine Learning – ICANN 2022. ICANN 2022. Lecture Notes in Computer Science, vol 13530. Springer, Cham. https://doi.org/10.1007/978-3-031-15931-2_31

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-15931-2_31

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-15930-5

  • Online ISBN: 978-3-031-15931-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics