Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Hydra: : Hybrid-model federated learning for human activity recognition on heterogeneous devices

Published: 17 April 2024 Publication History

Abstract

Federated Learning (FL) has recently received extensive attention in enabling privacy-preserving edge AI services for Human Activity Recognition (HAR). However, users’ mobile and wearable devices in the HAR scenario usually possess dramatically different computing capability and diverse data distributions, making it very challenging for such heterogeneous HAR devices to conduct effective collaborative training (co-training) with the traditional FL schemes. To address this issue, we present Hydra, a Hybrid-model federated learning mechanism that facilitates the co-training among heterogeneous devices by allowing them to train models that well fit their own computing capability. Specifically, Hydra leverages BranchyNet to design a large-small global hybrid-model and enables heterogeneous devices to train the proper parts of the model tailored to their computing capability. Hydra drives co-training among the devices and clusters them based on model similarity to mitigate the impact of HAR data heterogeneity on model accuracy. In order to deal with the issue that large model may lack sufficient training data due to the limited number of high-performance devices in FL, we introduce a pairing scheme between high and low performance devices for effective co-training, and further propose sample selection approach to select more valuable samples to participate in co-training. We then formulate a constrained co-training problem within a cluster that is proved to be NP-hard and devise a fast greedy-based heuristic algorithm to solve it. In addition, to address the low accuracy of small models, we also propose a Large-to-Small knowledge distillation algorithm for resource-constrained devices to optimize the efficiency of transferring knowledge from large models to small models. We conduct extensive experiments on three HAR datasets and the experimental results demonstrate the superior performance of Hydra for achieving outstanding model accuracy improvement compared with other state-of-the-art schemes.

References

[1]
X. Ouyang, X. Shuai, J. Zhou, I.W. Shi, Z. Xie, G. Xing, J. Huang, Cosmo: contrastive fusion learning with small data for multimodal human activity recognition, in: Proceedings of the 28th Annual International Conference on Mobile Computing and Networking, 2022, pp. 324–337.
[2]
Wu Q., Chen X., Zhou Z., Zhang J., Fedhome: Cloud-edge based personalized federated learning for in-home health monitoring, IEEE Trans. Mob. Comput. 21 (8) (2020) 2818–2832.
[3]
Ramasamy Ramamurthy S., Roy N., Recent trends in machine learning for human activity recognition—A survey, Wiley Interdiscip. Rev. Data Min. Knowl. Discovery 8 (4) (2018).
[4]
Gu F., Chung M.-H., Chignell M., Valaee S., Zhou B., Liu X., A survey on deep learning for human activity recognition, ACM Comput. Surv. 54 (8) (2021) 1–34.
[5]
Chen K., Zhang D., Yao L., Guo B., Yu Z., Liu Y., Deep learning for sensor-based human activity recognition: Overview, challenges, and opportunities, ACM Comput. Surv. 54 (4) (2021) 1–40.
[6]
Albrecht J.P., How the GDPR will change the world, Eur. Data Prot. L. Rev. 2 (2016) 287.
[7]
Konečnỳ J., McMahan H.B., Ramage D., Richtárik P., Federated optimization: Distributed machine learning for on-device intelligence, 2016, arXiv preprint arXiv:1610.02527.
[8]
Hard A., Rao K., Mathews R., Ramaswamy S., Beaufays F., Augenstein S., Eichner H., Kiddon C., Ramage D., Federated learning for mobile keyboard prediction, 2018, arXiv preprint arXiv:1811.03604.
[9]
Konečnỳ J., McMahan H.B., Yu F.X., Richtárik P., Suresh A.T., Bacon D., Federated learning: Strategies for improving communication efficiency, 2016, arXiv preprint arXiv:1610.05492.
[10]
Luo S., Chen X., Wu Q., Zhou Z., Yu S., HFEL: Joint edge association and resource allocation for cost-efficient hierarchical federated edge learning, IEEE Trans. Wireless Commun. 19 (10) (2020) 6535–6548.
[11]
Bonawitz K., Eichner H., Grieskamp W., Huba D., Ingerman A., Ivanov V., Kiddon C., Konečnỳ J., Mazzocchi S., McMahan B., et al., Towards federated learning at scale: System design, Proc. Mach. Learn. Syst. 1 (2019) 374–388.
[12]
C. Li, D. Niu, B. Jiang, X. Zuo, J. Yang, Meta-har: Federated representation learning for human activity recognition, in: Proceedings of the Web Conference 2021, 2021, pp. 912–922.
[13]
Presotto R., Civitarese G., Bettini C., FedCLAR: Federated clustering for personalized sensor-based human activity recognition, in: 2022 IEEE International Conference on Pervasive Computing and Communications, PerCom, IEEE, 2022, pp. 227–236.
[14]
Nishio T., Yonetani R., Client selection for federated learning with heterogeneous resources in mobile edge, in: ICC 2019-2019 IEEE International Conference on Communications, ICC, IEEE, 2019, pp. 1–7.
[15]
Sze V., Chen Y.-H., Yang T.-J., Emer J.S., Efficient processing of deep neural networks: A tutorial and survey, Proc. IEEE 105 (12) (2017) 2295–2329.
[16]
Bicheno S., Smartphone market headed for mid-range mediocrity, 2023, https://telecoms.com/506561/smartphone-market-headed-for-mid-range-mediocrity/. (Accessed 10 March 2023).
[17]
X. Ouyang, Z. Xie, J. Zhou, J. Huang, G. Xing, Clusterfl: a similarity-aware federated learning system for human activity recognition, in: Proceedings of the 19th Annual International Conference on Mobile Systems, Applications, and Services, 2021, pp. 54–66.
[18]
Zhao Y., Li M., Lai L., Suda N., Civin D., Chandra V., Federated learning with non-iid data, 2018, arXiv preprint arXiv:1806.00582.
[19]
Teerapittayanon S., McDanel B., Kung H.-T., Branchynet: Fast inference via early exiting from deep neural networks, in: 2016 23rd International Conference on Pattern Recognition, ICPR, IEEE, 2016, pp. 2464–2469.
[20]
Agarwal N., Suresh A.T., Yu F.X.X., Kumar S., McMahan B., Cpsgd: Communication-efficient and differentially-private distributed SGD, Adv. Neural Inf. Process. Syst. 31 (2018).
[21]
McMahan H.B., Ramage D., Talwar K., Zhang L., Learning differentially private recurrent language models, in: International Conference on Learning Representations, 2018, URL https://openreview.net/forum?id=BJ0hF1Z0b.
[22]
J. Wang, J. Zhang, W. Bao, X. Zhu, B. Cao, P.S. Yu, Not just privacy: Improving performance of private deep learning in mobile cloud, in: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2018, pp. 2407–2416.
[23]
Zhang X., Hu M., Xia J., Wei T., Chen M., Hu S., Efficient federated learning for cloud-based AIoT applications, IEEE Trans. Comput.-Aided Des. Integr. Circuits Syst. 40 (11) (2020) 2211–2223.
[24]
Yang L., Huang J., Lin W., Cao J., Personalized federated learning on non-IID data via group-based meta-learning, ACM Trans. Knowl. Discov. Data (TKDD) (2022).
[25]
N.D. Lane, Y. Xu, H. Lu, S. Hu, T. Choudhury, A.T. Campbell, F. Zhao, Enabling large-scale human activity inference on smartphones using community similarity networks (csn), in: Proceedings of the 13th International Conference on Ubiquitous Computing, 2011, pp. 355–364.
[26]
Smith V., Chiang C.-K., Sanjabi M., Talwalkar A.S., Federated multi-task learning, Adv. Neural Inf. Process. Syst. 30 (2017).
[27]
J. Shin, Y. Li, Y. Liu, S.-J. Lee, FedBalancer: data and pace control for efficient federated learning on heterogeneous clients, in: Proceedings of the 20th Annual International Conference on Mobile Systems, Applications and Services, 2022, pp. 436–449.
[28]
L. Nagalapatti, R.S. Mittal, R. Narayanam, Is your data relevant?: Dynamic selection of relevant data for federated learning, in: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 36, No. 7, 2022, pp. 7859–7867.
[29]
L. Zhang, J. Song, A. Gao, J. Chen, C. Bao, K. Ma, Be your own teacher: Improve the performance of convolutional neural networks via self distillation, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 3713–3722.
[30]
Xu C., Gao W., Li T., Bai N., Li G., Zhang Y., Teacher-student collaborative knowledge distillation for image classification, Appl. Intell. (2022) 1–13.
[31]
L. Tu, X. Ouyang, J. Zhou, Y. He, G. Xing, Feddl: Federated learning via dynamic layer sharing for human activity recognition, in: Proceedings of the 19th ACM Conference on Embedded Networked Sensor Systems, 2021, pp. 15–28.
[32]
Chen Y., Qin X., Wang J., Yu C., Gao W., FedHealth: A federated transfer learning framework for wearable healthcare, IEEE Intell. Syst. 35 (4) (2020) 83–93.
[33]
Yu H., Chen Z., Zhang X., Chen X., Zhuang F., Xiong H., Cheng X., FedHAR: Semi-supervised online learning for personalized federated human activity recognition, IEEE Trans. Mob. Comput. (2021) 1.
[34]
Mo F., Malekzadeh M., Chatterjee S., Kawsar F., Mathur A., Centaur: Federated learning for constrained edge devices, 2022.
[35]
Shi H., Radu V., Data selection for efficient model update in federated learning, in: Proceedings of the 2nd European Workshop on Machine Learning and Systems, EuroMLSys ’22, Association for Computing Machinery, New York, NY, USA, ISBN 9781450392549, 2022, pp. 72–78,.
[36]
Li A., Zhang L., Tan J., Qin Y., Wang J., Li X.-Y., Sample-level data selection for federated learning, in: IEEE INFOCOM 2021-IEEE Conference on Computer Communications, IEEE, 2021, pp. 1–10.
[37]
L. Zhang, L. Shen, L. Ding, D. Tao, L.-Y. Duan, Fine-tuning Global Model via Data-Free Knowledge Distillation for Non-IID Federated Learning, in: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR, 2022, pp. 10164–10173.
[38]
Itahara S., Nishio T., Koda Y., Morikura M., Yamamoto K., Distillation-based semi-supervised federated learning for communication-efficient co-training with non-IID private data, IEEE Trans. Mob. Comput. 22 (1) (2023) 191–205.
[39]
H.Q. Le, L.X. Nguyen, S.-B. Park, C.S. Hong, Layer-wise Knowledge Distillation for Cross-Device Federated Learning, in: 2023 International Conference on Information Networking, ICOIN, 2023, pp. 526–529.
[40]
Xing H., Xiao Z., Qu R., Zhu Z., Zhao B., An efficient federated distillation learning system for multitask time series classification, IEEE Trans. Instrum. Meas. 71 (2022) 1–12.
[41]
Tang S., Chen L., He K., Xia J., Fan L., Nallanathan A., Computational intelligence and deep learning for next-generation edge-enabled industrial IoT, IEEE Trans. Netw. Sci. Eng. (2022).
[42]
Xu C., Qu Y., Xiang Y., Gao L., Asynchronous federated learning on heterogeneous devices: A survey, 2021, arXiv preprint arXiv:2109.04269.
[43]
Imteaj A., Thakker U., Wang S., Li J., Amini M.H., A survey on federated learning for resource-constrained IoT devices, IEEE Internet Things J. 9 (1) (2021) 1–24.
[44]
Xie C., Koyejo S., Gupta I., Asynchronous federated optimization, 2019, arXiv preprint arXiv:1903.03934.
[45]
T. Zhang, L. Gao, S. Lee, M. Zhang, S. Avestimehr, TimelyFL: Heterogeneity-aware Asynchronous Federated Learning with Adaptive Partial Training, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 5063–5072.
[46]
Zhang Y., Liu D., Duan M., Li L., Chen X., Ren A., Tan Y., Wang C., FedMDS: An efficient model discrepancy-aware semi-asynchronous clustered federated learning framework, IEEE Trans. Parallel Distrib. Syst. 34 (3) (2023) 1007–1019.
[47]
Jacob L., Vert J.-p., Bach F., Clustered multi-task learning: A convex formulation, Adv. Neural Inf. Process. Syst. 21 (2008).
[48]
Evgeniou T., Micchelli C.A., Pontil M., Shawe-Taylor J., Learning multiple tasks with kernel methods., J. Mach. Learn. Res. 6 (4) (2005).
[49]
Kullback S., Information Theory and Statistics, Courier Corporation, 1997.
[50]
Hinton G., Vinyals O., Dean J., Distilling the knowledge in a neural network, 2015, arXiv preprint arXiv:1503.02531.
[51]
Wang H., Feng L., Meng X., Chen Z., Yu L., Zhang H., Multi-view metric learning based on KL-divergence for similarity measurement, Neurocomputing 238 (2017) 269–276.
[52]
C. Ding, X. He, K-means clustering via principal component analysis, in: Proceedings of the Twenty-First International Conference on Machine Learning, 2004, p. 29.
[53]
Katharopoulos A., Fleuret F., Not all samples are created equal: Deep learning with importance sampling, in: International Conference on Machine Learning, PMLR, 2018, pp. 2525–2534.
[54]
Johnson T.B., Guestrin C., Training deep models faster with robust, approximate importance sampling, Adv. Neural Inf. Process. Syst. 31 (2018).
[55]
F. Lai, X. Zhu, H.V. Madhyastha, M. Chowdhury, Oort: Efficient federated learning via guided participant selection, in: 15th {U S E N I X} Symposium on Operating Systems Design and Implementation ({O S D I} 21), 2021, pp. 19–35.
[56]
J. Böhler, F. Kerschbaum, Secure multi-party computation of differentially private median, in: Proceedings of the 29th USENIX Conference on Security Symposium, 2020, pp. 2147–2164.
[57]
Song H., Kim M., Park D., Shin Y., Lee J.-G., Learning from noisy labels with deep neural networks: A survey, IEEE Trans. Neural Netw. Learn. Syst. (2022).
[58]
Wang L., Yoon K.-J., Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks, IEEE Trans. Pattern Anal. Mach. Intell. (2021).
[59]
Caprara A., Kellerer H., Pferschy U., A PTAS for the multiple subset sum problem with different knapsack capacities, Inform. Process. Lett. 73 (3–4) (2000) 111–118.
[60]
Cho H., Mathur A., Kawsar F., Flame: Federated learning across multi-device environments, Proc. ACM Interact Mobile Wearable Ubiquitous Technol. 6 (3) (2022) 1–29.
[61]
J. Sun, A. Li, L. Duan, S. Alam, X. Deng, X. Guo, H. Wang, M. Gorlatova, M. Zhang, H. Li, et al., FedSEA: A Semi-Asynchronous Federated Learning Framework for Extremely Heterogeneous Devices, in: Proceedings of the 20th ACM Conference on Embedded Networked Sensor Systems, 2022, pp. 106–119.
[62]
Zeng L., Chen X., Zhou Z., Yang L., Zhang J., Coedge: Cooperative dnn inference with adaptive workload partitioning over heterogeneous edge devices, IEEE/ACM Trans. Netw. 29 (2) (2020) 595–608.
[63]
S. Ye, L. Zeng, Q. Wu, K. Luo, Q. Fang, X. Chen, Eco-FL: Adaptive Federated Learning with Efficient Edge Collaborative Pipeline Training, in: Proceedings of the 51st International Conference on Parallel Processing, 2022, pp. 1–11.
[64]
Zhong Z., Bao W., Wang J., Zhu X., Zhang X., Flee: A hierarchical federated learning framework for distributed deep neural network over cloud, edge, and end device, ACM Trans. Intell. Syst. Technol. 13 (5) (2022) 1–24.
[65]
A. Stisen, H. Blunck, S. Bhattacharya, T.S. Prentow, M.B. Kjærgaard, A. Dey, T. Sonne, M.M. Jensen, Smart devices are different: Assessing and mitigatingmobile sensing heterogeneities for activity recognition, in: Proceedings of the 13th ACM Conference on Embedded Networked Sensor Systems, 2015, pp. 127–140.
[66]
Vavoulas G., Chatzaki C., Malliotakis T., Pediaditis M., Tsiknakis M., The MobiAct dataset: Recognition of activities of daily living using smartphones., in: ICT4AgeingWell, Rome, 2016, pp. 143–151.
[67]
McMahan B., Moore E., Ramage D., Hampson S., y Arcas B.A., Communication-efficient learning of deep networks from decentralized data, in: Artificial Intelligence and Statistics, PMLR, 2017, pp. 1273–1282.
[68]
Jia B., Zhang X., Liu J., Zhang Y., Huang K., Liang Y., Blockchain-enabled federated learning data protection aggregation scheme with differential privacy and homomorphic encryption in IIoT, IEEE Trans. Ind. Inform. 18 (6) (2021) 4049–4058.
[69]
Zhang C., Li S., Xia J., Wang W., Yan F., Liu Y., {B a t c h C r y p t}: Efficient homomorphic encryption for {C r o s s − S i l o} federated learning, in: 2020 USENIX Annual Technical Conference, USENIX ATC 20, 2020, pp. 493–506.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Journal of Systems Architecture: the EUROMICRO Journal
Journal of Systems Architecture: the EUROMICRO Journal  Volume 147, Issue C
Feb 2024
138 pages

Publisher

Elsevier North-Holland, Inc.

United States

Publication History

Published: 17 April 2024

Author Tags

  1. Edge AI
  2. Federated learning
  3. Human activity recognition

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 0
    Total Downloads
  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 22 Sep 2024

Other Metrics

Citations

View Options

View options

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media