Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Multi-objective Learning to Overcome Catastrophic Forgetting in Time-series Applications

Published: 30 July 2022 Publication History

Abstract

One key objective of artificial intelligence involves the continuous adaptation of machine learning models to new tasks. This branch of continual learning is also referred to as lifelong learning (LL), where a major challenge is to minimize catastrophic forgetting, or forgetting previously learned tasks. While previous work on catastrophic forgetting has been focused on vision problems; this work targets time-series data. In addition to choosing an architecture appropriate for time-series sequences, our work addresses limitations in previous work, including the handling of distribution shifts in class labels. We present multi-objective learning with three loss functions to minimize catastrophic forgetting, prediction error, and errors in generalizing across label shifts, simultaneously. We build a multi-task autoencoder network with a hierarchical convolutional recurrent architecture. The proposed method is capable of learning multiple time-series tasks simultaneously. For cases where the model needs to learn multiple new tasks, we propose sequential learning, starting with tasks that have the best individual performances. This solution was evaluated on four benchmark human activity recognition datasets collected from mobile sensing devices. A wide set of baseline comparisons is performed, and an ablation analysis is run to evaluate the impact of the different losses in the proposed multi-objective method. The results demonstrate an up to 4% performance improvement in catastrophic forgetting compared to the use of loss functions in state-of-the-art solutions while demonstrating minimal losses compared to upper bound methods of traditional fine-tuning (FT) and multi-task learning (MTL).

References

[1]
James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A. Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, Demis Hassabis, Claudia Clopath, Dharshan Kumaran, and Raia Hadsell. 2017. Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences 114, 13 (2017), 3521–3526.
[2]
Matthias De Lange, Rahaf Aljundi, Marc Masana, Sarah Parisot, Xu Jia, Aleš Leonardis, Gregory Slabaugh, and Tinne Tuytelaars. 2019. A continual learning survey: Defying forgetting in classification tasks. IEEE Transactions on Pattern Analysis and Machine Intelligence 44, 7 (2019), 3366–3385.
[3]
Zhizhong Li and Derek Hoiem. 2018. Learning without forgetting. IEEE Transactions on Pattern Analysis and Machine Intelligence 40, 12 (2018), 2935–2947.
[4]
Amal Rannen, Rahaf Aljundi, Matthew B. Blaschko, and Tinne Tuytelaars. 2017. Encoder-based lifelong learning. In Proceedings of the IEEE International Conference on Computer Vision. 1320–1328.
[5]
Yuichi Hattori, Sozo Inoue, and Go Hirakawa. 2011. A large scale gathering system for activity data with mobile sensors. In Proceedings of the 2011 15th Annual International Symposium on Wearable Computers. IEEE, 97–100.
[6]
Billur Barshan and Murat Cihan Yüksek. 2014. Recognizing daily and sports activities in two open source machine learning environments using body-worn sensor units. The Computer Journal 57, 11 (2014), 1649–1667.
[7]
Ricardo Chavarriaga, Hesam Sagha, Alberto Calatroni, Sundara Tejaswi Digumarti, Gerhard Tröster, José del R. Millán, and Daniel Roggen. 2013. The Opportunity challenge: A benchmark database for on-body sensor-based activity recognition. Pattern Recognition Letters 34, 15 (2013), 2033–2042.
[8]
Davide Anguita, Alessandro Ghio, Luca Oneto, Xavier Parra, and Jorge Luis Reyes-Ortiz. 2013. A public domain dataset for human activity recognition using smartphones. In Proceedings of the 21st International European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning.
[9]
Hassan Ismail Fawaz, Germain Forestier, Jonathan Weber, Lhassane Idoumghar, and Pierre-Alain Muller. 2018. Transfer learning for time series classification. In Proceedings of the 2018 IEEE International Conference on Big Data. IEEE, 1367–1376.
[10]
Rich Caruana. 1997. Multi-task learning. Machine Learning 28, 1 (1997), 41–75.
[11]
Yu Zhang and Dit-Yan Yeung. 2014. A regularization approach to learning task relationships in multitask learning. ACM Transactions on Knowledge Discovery from Data 8, 3 (2014), 1–31.
[12]
Reem A. Mahmoud, Hazem Hajj, and Fadi N. Karameh. 2020. A systematic approach to multi-task learning from time-series data. Applied Soft Computing 96 (2020), 106586.
[13]
Xinjiang Lu, Zhiwen Yu, Chuanren Liu, Yanchi Liu, Hui Xiong, and Bin Guo. 2020. Inferring lifetime status of point-of-interest: A multitask multiclass approach. ACM Transactions on Knowledge Discovery from Data 14, 1 (2020), 1–27.
[14]
Anthony Robins. 1995. Catastrophic forgetting, rehearsal and pseudorehearsal. Connection Science 7, 2 (1995), 123–146.
[15]
Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems. 2672–2680.
[16]
Hanul Shin, Jung Kwon Lee, Jaehong Kim, and Jiwon Kim. 2017. Continual learning with deep generative replay. In Proceedings of the Advances in Neural Information Processing Systems. 2990–2999.
[17]
Ju Xu and Zhanxing Zhu. 2018. Reinforced continual learning. In Proceedings of the Advances in Neural Information Processing Systems. 899–908.
[18]
Andrei A. Rusu, Neil C. Rabinowitz, Guillaume Desjardins, Hubert Soyer, James Kirkpatrick, Koray Kavukcuoglu, Razvan Pascanu, and Raia Hadsell. 2016. Progressive neural networks. arXiv preprint arXiv:1606.04671.
[19]
Chrisantha Fernando, Dylan Banarse, Charles Blundell, Yori Zwols, David Ha, Andrei A. Rusu, Alexander Pritzel, and Daan Wierstra. 2017. Pathnet: Evolution channels gradient descent in super neural networks. arXiv preprint arXiv:1701.08734.
[20]
Joan Serra, Didac Suris, Marius Miron, and Alexandros Karatzoglou. 2018. Overcoming catastrophic forgetting with hard attention to the task. In International Conference on Machine Learning. PMLR, 4548–4557.
[21]
Benedikt Pfülb and Alexander Gepperth. 2019. A comprehensive, application-oriented study of catastrophic forgetting in dnns. arXiv:1905.08101. Retrieved from https://arxiv.org/abs/1905.08101.
[22]
Friedemann Zenke, Ben Poole, and Surya Ganguli. 2017. Continual learning through synaptic intelligence. Proceedings of Machine Learning Research 70, 3995 (2017), 3987.
[23]
Sebastian Farquhar and Yarin Gal. 2018. Towards robust evaluations of continual learning. arXiv:1805.09733. Retrieved from https://arxiv.org/abs/1805.09733.
[24]
Qi Wang, Liang Zhan, Paul Thompson, and Jiayu Zhou. 2020. Multimodal learning with incomplete modalities by knowledge distillation. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 1828–1838.
[25]
Benjamin Lindemann, Timo Müller, Hannes Vietz, Nasser Jazdi, and Michael Weyrich. 2021. A survey on long short-term memory networks for time series prediction. Procedia CIRP 99 (2021), 650–655.
[26]
Jianbo Yang, Minh Nhut Nguyen, Phyo Phyo San, Xiaoli Li, and Shonali Krishnaswamy. 2015. Deep convolutional neural networks on multichannel time series for human activity recognition. In Proceedings of the 24th International Joint Conference on Artificial Intelligence. 3995–4001.
[27]
Charissa Ann Ronao and Sung-Bae Cho. 2016. Human activity recognition with smartphone sensors using deep learning neural networks. Expert Systems with Applications 59 (2016), 235–244.
[28]
Robert D. Chambers and Nathanael C. Yoder. 2020. FilterNet: A many-to-many deep learning architecture for time series classification. Sensors 20, 9 (2020), 2498.
[29]
Shuochao Yao, Shaohan Hu, Yiran Zhao, Aston Zhang, and Tarek Abdelzaher. 2017. Deepsense: A unified deep learning framework for time-series mobile sensing data processing. In Proceedings of the 26th International Conference on World Wide Web. 351–360.
[30]
Jia Wang, Tong Sun, Benyuan Liu, Yu Cao, and Hongwei Zhu. 2021. CLVSA: A convolutional LSTM based variational sequence-to-sequence model with attention for predicting trends of financial markets. arXiv preprint arXiv:2104.04041.
[31]
G. Hinton, O. Vinyals, and J. Dean. 2014. Distilling the knowledge in a neural network. NIPS Deep Learning Workshop. arXiv preprint arXiv:1503.02531 2, no. 7.
[32]
Zhiqiang Gao, Dawei Liu, Kaizhu Huang, and Yi Huang. 2019. Context-aware human activity and smartphone position-mining with motion sensors. Remote Sensing 11, 21 (2019), 2531.
[33]
Andrew Ng. 2011. Sparse autoencoder. CS294A Lecture Notes 72, 2011 (2011), 1–19.
[34]
Jindong Wang, Yiqiang Chen, Shuji Hao, Xiaohui Peng, and Lisha Hu. 2019. Deep learning for sensor-based activity recognition: A survey. Pattern Recognition Letters 119 (2019), 3–11.
[35]
Xuezhi Wang and Jeff G. Schneider. 2015. Generalization bounds for transfer learning under model shift. In Proceedings of the UAI. 922–931.

Cited By

View all
  • (2024)Comprehensive Sensitivity Analysis Framework for Transfer Learning Performance Assessment for Time Series Forecasting: Basic Concepts and Selected Case StudiesSymmetry10.3390/sym1602024116:2(241)Online publication date: 16-Feb-2024
  • (2024)Multi-Task Learning with Sequential Dependence Toward Industrial Applications: A Systematic FormulationACM Transactions on Knowledge Discovery from Data10.1145/364046818:5(1-29)Online publication date: 28-Feb-2024
  • (2024)A Novel Three-Way Deep Learning Approach for Multigranularity Fuzzy Association Analysis of Time Series DataIEEE Transactions on Fuzzy Systems10.1109/TFUZZ.2023.333292132:9(4835-4845)Online publication date: 1-Sep-2024
  • Show More Cited By

Index Terms

  1. Multi-objective Learning to Overcome Catastrophic Forgetting in Time-series Applications

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Knowledge Discovery from Data
    ACM Transactions on Knowledge Discovery from Data  Volume 16, Issue 6
    December 2022
    631 pages
    ISSN:1556-4681
    EISSN:1556-472X
    DOI:10.1145/3543989
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 30 July 2022
    Online AM: 17 June 2022
    Accepted: 01 November 2021
    Revised: 01 May 2021
    Received: 01 November 2020
    Published in TKDD Volume 16, Issue 6

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Time-series
    2. knowledge distillation
    3. multitask learning
    4. autoencoders

    Qualifiers

    • Research-article
    • Refereed

    Funding Sources

    • American University of Beirut’s University Research Board (AUB URB)

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)106
    • Downloads (Last 6 weeks)7
    Reflects downloads up to 24 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Comprehensive Sensitivity Analysis Framework for Transfer Learning Performance Assessment for Time Series Forecasting: Basic Concepts and Selected Case StudiesSymmetry10.3390/sym1602024116:2(241)Online publication date: 16-Feb-2024
    • (2024)Multi-Task Learning with Sequential Dependence Toward Industrial Applications: A Systematic FormulationACM Transactions on Knowledge Discovery from Data10.1145/364046818:5(1-29)Online publication date: 28-Feb-2024
    • (2024)A Novel Three-Way Deep Learning Approach for Multigranularity Fuzzy Association Analysis of Time Series DataIEEE Transactions on Fuzzy Systems10.1109/TFUZZ.2023.333292132:9(4835-4845)Online publication date: 1-Sep-2024
    • (2024)Resource-Efficient Continual Learning for Personalized Online Seizure Detection2024 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)10.1109/EMBC53108.2024.10781699(1-7)Online publication date: 15-Jul-2024
    • (2024)Fault diagnosis of pressure relief valve based on improved deep Residual Shrinking NetworkMeasurement10.1016/j.measurement.2023.113752224(113752)Online publication date: Jan-2024
    • (2024)Knowledge transfer in lifelong machine learning: a systematic literature reviewArtificial Intelligence Review10.1007/s10462-024-10853-957:8Online publication date: 26-Jul-2024

    View Options

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media