Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

FLEE: A Hierarchical Federated Learning Framework for Distributed Deep Neural Network over Cloud, Edge, and End Device

Published: 13 October 2022 Publication History

Abstract

With the development of smart devices, the computing capabilities of portable end devices such as mobile phones have been greatly enhanced. Meanwhile, traditional cloud computing faces great challenges caused by privacy-leakage and time-delay problems, there is a trend to push models down to edges and end devices. However, due to the limitation of computing resource, it is difficult for end devices to complete complex computing tasks alone. Therefore, this article divides the model into two parts and deploys them on multiple end devices and edges, respectively. Meanwhile, an early exit is set to reduce computing resource overhead, forming a hierarchical distributed architecture. In order to enable the distributed model to continuously evolve by using new data generated by end devices, we comprehensively consider various data distributions on end devices and edges, proposing a hierarchical federated learning framework FLEE, which can realize dynamical updates of models without redeploying them. Through image and sentence classification experiments, we verify that it can improve model performances under all kinds of data distributions, and prove that compared with other frameworks, the models trained by FLEE consume less global computing resource in the inference stage.

Appendix

Algorithm FedAvg illustrates the process of aggregation, its inputs include the number of clients \(M\), local iteration times \(E\), training batch size \(B\), learning rate \(\eta\), and total aggregation times \(K\). In this algorithm, firstly, each client downloads the initialized model \(G_0\) from the server. Next, models are locally trained \(E\) times with \(CLIENTUPDATE\), and then sent to the server for aggregation. After that, the server sums up all the models according to the data proportion of each client, and finally distributes the aggregated model \(G^{j+1}\) to clients for further training. The process above iterates \(K\) times. It is worth noting that only a part of clients participate in the training process in each iteration, and the whole training process adopts the gradient descent method.

References

[1]
M. S. H. Abad, E. Ozfatura, D. GUndUz, and O. Ercetin. 2020. Hierarchical federated learning ACROSS heterogeneous cellular networks. In Proceedings of the 2020 IEEE International Conference on Acoustics, Speech, and Signal Processing. 8866–8870. DOI:
[2]
Manoj Alwani, Han Chen, Michael Ferdman, and Peter Milder. 2016. Fused-layer CNN accelerators. In Proceedings of the 2016 49th Annual IEEE/ACM International Symposium on Microarchitecture. 1–12. DOI:
[3]
Christopher Briggs, Zhong Fan, and Peter Andras. 2020. Federated learning with hierarchical clustering of local updates to improve training on non-IID data. In Proceedings of the 2020 International Joint Conference on Neural Networks. 1–9. DOI:
[4]
Wan-Jung Chang, Liang-Bi Chen, Yang-Kun Ou, Aaron Raymond See, and Tzu-Chin Yang. 2021. A bed-exit and bedside fall warning system based on deep learning edge computing techniques. In Proceedings of the 2021 IEEE 3rd Global Conference on Life Sciences and Technologies. 142–143. DOI:
[5]
Mingqing Chen, Rajiv Mathews, Tom Ouyang, and Françoise Beaufays. 2019. Federated Learning Of Out-Of-Vocabulary Words. (2019). arXiv:cs.CL/1903.10635. Retrieved from https://arxiv.org/abs/1903.10635.
[6]
Ming-Chuan Chiu, Jih-Hung Huang, Saraj Gupta, and Gulsen Akman. 2021. Developing a personalized recommendation system in a smart product service system based on unsupervised learning model. Computers in Industry 128, 1 (2021), 103421.
[7]
Marika Cusick, Prakash Adekkanattu, Thomas R. Campion, Evan T. Sholle, Annie Myers, Samprit Banerjee, George Alexopoulos, Yanshan Wang, and Jyotishman Pathak. 2021. Using weak supervision and deep learning to classify clinical notes for identification of current suicidal ideation. Journal of Psychiatric Research 136, 1 (2021), 95–102.
[8]
Maha Elbayad, Jiatao Gu, Edouard Grave, and Michael Auli. 2020. Depth-Adaptive Transformer. In International Conference on Learning Representations. https://openreview.net/forum?id=SJg7KhVKPH
[9]
Dashan Gao, Ce Ju, Xiguang Wei, Yang Liu, Tianjian Chen, and Qiang Yang. 2020. HHHFL: Hierarchical Heterogeneous Horizontal Federated Learning for Electroencephalography. (2020). arXiv:eess.SP/1909.05784. Retrieved from https://arxiv.org/abs/1909.05784.
[10]
Seyyedali Hosseinalipour, Christopher G. Brinton, Vaneet Aggarwal, Huaiyu Dai, and Mung Chiang. 2020. From federated to fog learning: Distributed machine learning over heterogeneous wireless networks. IEEE Communications Magazine 58, 12 (2020), 41–47. DOI:
[11]
Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. (2017). arXiv:cs.CV/1704.04861. Retrieved from https://arxiv.org/abs/1704.04861.
[12]
Maryam Imani. 2021. Automatic diagnosis of coronavirus (COVID-19) using shape and texture characteristics extracted from X-Ray and CT-Scan images. Biomedical Signal Processing and Control 68, 1 (2021), 102602.
[13]
Yiping Kang, Johann Hauswald, Cao Gao, Austin Rovinski, Trevor Mudge, Jason Mars, and Lingjia Tang. 2017. Neurosurgeon: Collaborative intelligence between the cloud and mobile edge. ACM SIGPLAN Notices 45, 1 (April 2017), 615–629. DOI:
[14]
Yoon Kim. 2014. Convolutional Neural Networks for Sentence Classification. (2014). arXiv:cs.CL/1408.5882. Retrieved from https://arxiv.org/abs/1408.5882.
[15]
Jakub Konečný, H. Brendan McMahan, Felix X. Yu, Peter Richtárik, Ananda Theertha Suresh, and Dave Bacon. 2017. Federated Learning: Strategies for Improving Communication Efficiency. (2017). arXiv:cs.LG/1610.05492. Retrieved from https://arxiv.org/abs/1610.05492.
[16]
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2017. ImageNet classification with deep convolutional neural networks. Communication of the ACM 60, 6 (May 2017), 84–90. DOI:
[17]
Hao Li, Hong Zhang, Xiaojuan Qi, Ruigang Yang, and Gao Huang. 2019. Improved techniques for training adaptive deep networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision.
[18]
Xiang Li, Kaixuan Huang, Wenhao Yang, Shusen Wang, and Zhihua Zhang. 2020. On the Convergence of FedAvg on Non-IID Data. In International Conference on Learning Representations. https://openreview.net/forum?id=HJxNAnVtDS
[19]
Kaiyuan Liao, Yi Zhang, Xuancheng Ren, Qi Su, Xu Sun, and Bin He. 2021. A global past-future early exit method for accelerating inference of pre-trained language models. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, Online, 2013–2023. DOI:
[20]
Lumin Liu, Jun Zhang, S. H. Song, and Khaled B. Letaief. 2020. Client-edge-cloud hierarchical federated learning. In Proceedings of the 2020 IEEE International Conference on Communications. 1–6. DOI:
[21]
Shuo Liu, Mingliang Gao, Vijay John, Zheng Liu, and Erik Blasch. 2020. Deep learning thermal image translation for night vision perception. ACM Transactions on Intelligent Systems and Technology 12, 1 (Dec. 2020), Article 9, 18 pages. DOI:
[22]
Weijie Liu, Peng Zhou, Zhiruo Wang, Zhe Zhao, Haotang Deng, and Qi Ju. 2020. FastBERT: a Self-distilling BERT with Adaptive Inference Time. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 6035–6044.
[23]
Guangxin Lou and Hongzhen Shi. 2020. Face image recognition based on convolutional neural network. China Communications 17, 2 (2020), 117–124. DOI:
[24]
Yoshitomo Matsubara and Marco Levorato. 2021. Neural compression and filtering for edge-assisted real-time object detection in challenged networks. In Proceedings of the 2020 25th International Conference on Pattern Recognition. IEEE, 2272–2279.
[25]
Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas. 2017. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (Proceedings of Machine Learning Research), Aarti Singh and Jerry Zhu (Eds.), Vol. 54. PMLR, 1273–1282. Retrieved from https://proceedings.mlr.press/v54/mcmahan17a.html.
[26]
Seyed Yahya Nikouei, Yu Chen, Sejun Song, Ronghua Xu, Baek-Young Choi, and Timothy Faughnan. 2018. Smart surveillance as an edge network service: From harr-cascade, SVM to a lightweight CNN. In Proceedings of the 2018 IEEE 4th International Conference on Collaboration and Internet Computing. 256–265. DOI:
[27]
Zhaolong Ning, Peiran Dong, Xiaojie Wang, Joel J. P. C. Rodrigues, and Feng Xia. 2019. Deep reinforcement learning for vehicular edge computing: An intelligent offloading system. ACM Transactions on Intelligent Systems and Technology 10, 6 (Oct. 2019), Article 60, 24 pages. DOI:
[28]
Hao Peng, Haoran Li, Yangqiu Song, Vincent Zheng, and Jianxin Li. 2021. Differentially Private Federated Knowledge Graphs Embedding. Association for Computing Machinery, New York, NY, 1416–1425. DOI:
[29]
Hao Peng, Renyu Yang, Zheng Wang, Jianxin Li, Lifang He, Philip Yu, Albert Zomaya, and Raj Ranjan. 2021. LIME: Low-cost incremental learning for dynamic heterogeneous information networks. IEEE Transactions on Computers 71, 3 (2021), 1–1. DOI:
[30]
Hao Peng, Ruitong Zhang, Yingtong Dou, Renyu Yang, Jingyi Zhang, and Philip S. Yu. 2021. Reinforced neighborhood selection guided multi-relational graph neural networks. ACM Transactions on Information Systems 40, 4 (Dec 2021), Article 69, 46 pages. DOI:
[31]
Mary Phuong and Christoph H. Lampert. 2019. Distillation-based training for multi-exit architectures. In Proceedings of the IEEE/CVF International Conference on Computer Vision.
[32]
Mohammad Rastegari, Vicente Ordonez, Joseph Redmon, and Ali Farhadi. 2016. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks. In Computer Vision – ECCV 2016, Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling (Eds.). Springer International Publishing, Cham, 525–542.
[33]
Felix Sattler, Simon Wiedemann, Klaus-Robert Müller, and Wojciech Samek. 2020. Robust and communication-efficient federated learning from non-i.i.d. data. IEEE Transactions on Neural Networks and Learning Systems 31, 9 (2020), 3400–3413. DOI:
[34]
Ragini Sharma, Saman Biookaghazadeh, Baoxin Li, and Ming Zhao. 2018. Are existing knowledge transfer techniques effective for deep learning with edge devices?. In Proceedings of the 2018 IEEE International Conference on Edge Computing. 42–49. DOI:
[35]
Karolj Skala, Davor Davidovic, Enis Afgan, Ivan Sovic, and Zorislav Sojat. 2015. Scalable distributed computing hierarchy: Cloud, fog and dew computing. Open Journal of Cloud Computing 2, 1 (2015), 16–24. Retrieved from http://nbn-resolving.de/urn:nbn:de:101:1-201705194519.
[36]
Luca Soldaini and Alessandro Moschitti. 2020. The Cascade Transformer: an Application for Efficient Answer Sentence Selection. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 5697–5708.
[37]
Surat Teerapittayanon, Bradley McDanel, and H. T. Kung. 2016. BranchyNet: Fast inference via early exiting from deep neural networks. In Proceedings of the 2016 23rd International Conference on Pattern Recognition. 2464–2469. DOI:
[38]
Surat Teerapittayanon, Bradley McDanel, and H. T. Kung. 2017. Distributed deep neural networks over the cloud, the edge and end devices. In Proceedings of the 2017 IEEE 37th International Conference on Distributed Computing Systems. 328–339. DOI:
[39]
Nguyen H. Tran, Wei Bao, Albert Zomaya, Minh N. H. Nguyen, and Choong Seon Hong. 2019. Federated learning over wireless networks: Optimization model design and analysis. In Proceedings of the IEEE INFOCOM 2019—IEEE Conference on Computer Communications. 1387–1395. DOI:
[40]
Aidmar Wainakh, Alejandro Sanchez Guinea, Tim Grube, and Max Mühlhäuser. 2020. Enhancing privacy via hierarchical federated learning. In Proceedings of the 2020 IEEE European Symposium on Security and Privacy Workshops. 344–347. DOI:
[41]
Dandan Wang, Changying Li, Huaibo Song, Hongting Xiong, Chang Liu, and Dongjian He. 2020. Deep Learning Approach for Apple Edge Detection to Remotely Monitor Apple Growth in Orchards. IEEE Access 8, 1 (2020), 26911–26925.
[42]
Ji Wang, Weidong Bao, Lichao Sun, Xiaomin Zhu, Bokai Cao, and Philip S. Yu. 2019. Private Model Compression via Knowledge Distillation. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence and Thirty-First Innovative Applications of Artificial Intelligence Conference and Ninth AAAI Symposium on Educational Advances in Artificial Intelligence (Honolulu, Hawaii, USA) (AAAI’19/IAAI’19/EAAI’19). AAAI Press, Article 147, 9 pages.
[43]
Jianyu Wang and Gauri Joshi. 2019. Adaptive Communication Strategies to Achieve the Best Error-Runtime Trade-off in Local-Update SGD. In Proceedings of Machine Learning and Systems, A. Talwalkar, V. Smith, and M. Zaharia (Eds.), Vol. 1. 212–229. https://proceedings.mlsys.org/paper/2019/file/c8ffe9a587b126f152ed3d89a146b445-Paper.pdf.
[44]
Jianyu Wang and Gauri Joshi. 2021. Cooperative SGD: A Unified Framework for the Design and Analysis of Local Update SGD Algorithms. Journal of Machine Learning Research 22, 213 (2021), 1–50. http://jmlr.org/papers/v22/20-147.html.
[45]
Shiqiang Wang, Tiffany Tuor, Theodoros Salonidis, Kin K. Leung, Christian Makaya, Ting He, and Kevin Chan. 2019. Adaptive federated learning in resource constrained edge computing systems. IEEE Journal on Selected Areas in Communications 37, 6 (2019), 1205–1221. DOI:
[46]
Xiaofei Wang, Yiwen Han, Victor C. M. Leung, Dusit Niyato, Xueqiang Yan, and Xu Chen. 2020. Convergence of edge computing and deep learning: A comprehensive survey. IEEE Communications Surveys Tutorials 22, 2 (2020), 869–904. DOI:
[47]
Yue Wang, Jianghao Shen, Ting-Kuei Hu, Pengfei Xu, Tan Nguyen, Richard Baraniuk, Zhangyang Wang, and Yingyan Lin. 2020. Dual dynamic inference: Enabling more efficient, adaptive, and controllable deep inference. IEEE Journal of Selected Topics in Signal Processing 14, 4 (2020), 623–633. DOI:
[48]
Xiaohang Xu, Hao Peng, Md Zakirul Alam Bhuiyan, Zhifeng Hao, Lianzhong Liu, Lichao Sun, and Lifang He. 2021. Privacy-preserving federated depression detection from multi-source mobile health data. IEEE Transactions on Industrial Informatics 18, 7 (2021), 4788–4797.
[49]
Bishan Yang, Scott Wen-tau Yih, Xiaodong He, Jianfeng Gao, and Li Deng. 2015. Embedding Entities and Relations for Learning and Inference in Knowledge Bases. In Proceedings of the International Conference on Learning Representations (ICLR) 2015. https://www.microsoft.com/en-us/research/publication/embedding-entities-and-relations-for-learning-and-inference-in-knowledge-bases/.
[50]
Le Yang, Yizeng Han, Xi Chen, Shiji Song, Jifeng Dai, and Gao Huang. 2020. Resolution adaptive networks for efficient inference. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[51]
Qiang Yang, Yang Liu, Tianjian Chen, and Yongxin Tong. 2019. Federated machine learning: Concept and applications. ACM Transactions on Intelligent Systems and Technology 10, 2 (Jan. 2019), Article 12, 19 pages. DOI:
[52]
Shanhe Yi, Zijiang Hao, Zhengrui Qin, and Qun Li. 2015. Fog computing: Platform and applications. In Proceedings of the 2015 3rd IEEE Workshop on Hot Topics in Web Systems and Technologies. 73–78. DOI:
[53]
Xiongtao Zhang, Xiaomin Zhu, Ji Wang, Hui Yan, Huangke Chen, and Weidong Bao. 2020. Federated learning with adaptive communication compression under dynamic bandwidth and unreliable networks. Information Sciences 540, 1 (2020), 242–262.
[54]
Zhuoran Zhao, Kamyar Mirzazad Barijough, and Andreas Gerstlauer. 2018. DeepThings: Distributed adaptive deep learning inference on resource-constrained IoT edge clusters. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems 37, 11 (2018), 2348–2359. DOI:
[55]
Zhihe Zhao, Zhehao Jiang, Neiwen Ling, Xian Shuai, and Guoliang Xing. 2018. ECRT: An edge computing system for real-time image-based object tracking. In Proceedings of the 16th ACM Conference on Embedded Networked Sensor Systems. Association for Computing Machinery, New York, NY, 394–395. DOI:
[56]
Wangchunshu Zhou, Canwen Xu, Tao Ge, Julian McAuley, Ke Xu, and Furu Wei. 2020. BERT Loses Patience: Fast and Robust Inference with Early Exit. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 18330–18341. https://proceedings.neurips.cc/paper/2020/file/d4dd111a4fd973394238aca5c05bebe3-Paper.pdf.
[57]
Hasib Zunair and A. Ben Hamza. 2021. Synthesis of COVID-19 chest X-rays using unpaired image-to-image translation. Social Network Analysis and Mining 11, 1 (2021), 23. DOI:

Cited By

View all
  • (2024)ATHENA-FL: Avoiding Statistical Heterogeneity with One-versus-All in Federated LearningJournal of Internet Services and Applications10.5753/jisa.2024.382615:1(273-288)Online publication date: 14-Aug-2024
  • (2024)Advancements in Federated Learning: Models, Methods, and PrivacyACM Computing Surveys10.1145/3664650Online publication date: 1-Jun-2024
  • (2024)Low-Latency Hierarchical Federated Learning in Wireless Edge NetworksIEEE Internet of Things Journal10.1109/JIOT.2023.331474311:4(6943-6960)Online publication date: 15-Feb-2024
  • Show More Cited By

Index Terms

  1. FLEE: A Hierarchical Federated Learning Framework for Distributed Deep Neural Network over Cloud, Edge, and End Device

        Recommendations

        Comments

        Information & Contributors

        Information

        Published In

        cover image ACM Transactions on Intelligent Systems and Technology
        ACM Transactions on Intelligent Systems and Technology  Volume 13, Issue 5
        October 2022
        424 pages
        ISSN:2157-6904
        EISSN:2157-6912
        DOI:10.1145/3542930
        • Editor:
        • Huan Liu
        Issue’s Table of Contents

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        Published: 13 October 2022
        Online AM: 17 May 2022
        Accepted: 30 January 2022
        Revised: 25 January 2022
        Received: 01 May 2021
        Published in TIST Volume 13, Issue 5

        Permissions

        Request permissions for this article.

        Check for updates

        Author Tags

        1. Federated learning
        2. early exit of inference
        3. distributed neural network

        Qualifiers

        • Research-article
        • Refereed

        Funding Sources

        • National Natural Science Foundation of China
        • Scientific Research Project of National University of Defense Technology

        Contributors

        Other Metrics

        Bibliometrics & Citations

        Bibliometrics

        Article Metrics

        • Downloads (Last 12 months)316
        • Downloads (Last 6 weeks)27
        Reflects downloads up to 23 Feb 2025

        Other Metrics

        Citations

        Cited By

        View all
        • (2024)ATHENA-FL: Avoiding Statistical Heterogeneity with One-versus-All in Federated LearningJournal of Internet Services and Applications10.5753/jisa.2024.382615:1(273-288)Online publication date: 14-Aug-2024
        • (2024)Advancements in Federated Learning: Models, Methods, and PrivacyACM Computing Surveys10.1145/3664650Online publication date: 1-Jun-2024
        • (2024)Low-Latency Hierarchical Federated Learning in Wireless Edge NetworksIEEE Internet of Things Journal10.1109/JIOT.2023.331474311:4(6943-6960)Online publication date: 15-Feb-2024
        • (2024)QuAsyncFL: Asynchronous Federated Learning With Quantization for Cloud–Edge–Terminal Collaboration Enabled AIoTIEEE Internet of Things Journal10.1109/JIOT.2023.329081811:1(59-69)Online publication date: 1-Jan-2024
        • (2024)Toward a Scalable and Energy-Efficient Framework for Industrial Cloud-Edge-IoT ContinuumIEEE Internet of Things Magazine10.1109/IOTM.001.23002297:5(14-20)Online publication date: Sep-2024
        • (2024)HydraJournal of Systems Architecture: the EUROMICRO Journal10.1016/j.sysarc.2023.103052147:COnline publication date: 17-Apr-2024
        • (2024)Federated fusion learning with attention mechanism for multi-client medical image analysisInformation Fusion10.1016/j.inffus.2024.102364108(102364)Online publication date: Aug-2024
        • (2024)Streaming traffic classification: a hybrid deep learning and big data approachCluster Computing10.1007/s10586-023-04234-027:4(5165-5193)Online publication date: 18-Jan-2024
        • (2023)Deep neural networks in the cloudNeurocomputing10.1016/j.neucom.2023.126327545:COnline publication date: 7-Aug-2023

        View Options

        Login options

        Full Access

        View options

        PDF

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        Full Text

        View this article in Full Text.

        Full Text

        HTML Format

        View this article in HTML Format.

        HTML Format

        Figures

        Tables

        Media

        Share

        Share

        Share this Publication link

        Share on social media