skip to main content
10.1145/3580305.3599521acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article
Free access

Theoretical Convergence Guaranteed Resource-Adaptive Federated Learning with Mixed Heterogeneity

Published: 04 August 2023 Publication History

Abstract

In this paper, we propose an adaptive learning paradigm for resource-constrained cross-device federated learning, in which heterogeneous local submodels with varying resources can be jointly trained to produce a global model. Different from existing studies, the submodel structures of different clients are formed by arbitrarily assigned neurons according to their local resources. Along this line, we first design a general resource-adaptive federated learning algorithm, namely RA-Fed, and rigorously prove its convergence with asymptotically optimal rate O(1/√Γ*TQ) under loose assumptions. Furthermore, to address both submodels heterogeneity and data heterogeneity challenges under non-uniform training, we come up with a new server aggregation mechanism RAM-Fed with the same theoretically proved convergence rate. Moreover, we shed light on several key factors impacting convergence, such as minimum coverage rate, data heterogeneity level, submodel induced noises. Finally, we conduct extensive experiments on two types of tasks with three widely used datasets under different experimental settings. Compared with the state-of-the-arts, our methods improve the accuracy up to 10% on average. Particularly, when submodels jointly train with 50% parameters, RAM-Fed achieves comparable accuracy to FedAvg trained with the full model.

Supplementary Material

MP4 File (rtfp1352-2min-promo.mp4)
In real-world cross-device federated learning scenarios, mobile devices are usually equipped with limited resources for computation and communication which seriously restrict the convergence performance of the federated learning algorithms. It would be difficult and unaffordable for the resource-constrained clients to run the full model for coordination in federated learning, especially for the arising large models like ChatGPT. Moreover, different clients have varying local resources. Therefore, it is essential to consider a novel learning paradigm in resource-limited federated learning in which different clients can train different submodels according to their own resource constraints. In addition, we consider more general cases without strong assumptions where existing works would become special cases of our proposed learning paradigm.
MP4 File (rtfp1352-20min-video.mp4)
In real-world cross-device federated learning scenarios, mobile devices are usually equipped with limited resources for computation and communication which seriously restrict the convergence performance of the federated learning algorithms. It would be difficult and unaffordable for the resource-constrained clients to run the full model for coordination in federated learning, especially for the arising large models like ChatGPT. Moreover, different clients have varying local resources. Therefore, it is essential to consider a novel learning paradigm in resource-limited federated learning in which different clients can train different submodels according to their own resource constraints. In addition, we consider more general cases without strong assumptions where existing works would become special cases of our proposed learning paradigm.

References

[1]
Enmao Diao, Jie Ding, and Vahid Tarokh. 2021. HeteroFL: Computation and Communication Efficient Federated Learning for Heterogeneous Clients. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net. https://openreview.net/forum?id=TNkPBBYFkXg
[2]
Xinran Gu, Kaixuan Huang, Jingzhao Zhang, and Longbo Huang. 2021. Fast Federated Learning in the Presence of Arbitrary Device Unavailability. In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual, Marc'Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (Eds.). 12052--12064. https://proceedings.neurips.cc/paper/2021/hash/64be20f6dd1dd46adf110cf871e3ed35-Abstract.html
[3]
Tzu-Ming Harry Hsu, Hang Qi, and Matthew Brown. 2019. Measuring the Effects of Non-Identical Data Distribution for Federated Visual Classification. CoRR, Vol. abs/1909.06335 (2019). [arXiv]1909.06335 http://arxiv.org/abs/1909.06335
[4]
Divyansh Jhunjhunwala, Pranay Sharma, Aushim Nagarkatti, and Gauri Joshi. 2022. Fedvarp: Tackling the variance due to partial client participation in federated learning. In Uncertainty in Artificial Intelligence, Proceedings of the Thirty-Eighth Conference on Uncertainty in Artificial Intelligence, UAI 2022, 1-5 August 2022, Eindhoven, The Netherlands (Proceedings of Machine Learning Research, Vol. 180), James Cussens and Kun Zhang (Eds.). PMLR, 906--916. https://proceedings.mlr.press/v180/jhunjhunwala22a.html
[5]
Yuang Jiang, Shiqiang Wang, Bong Jun Ko, Wei-Han Lee, and Leandros Tassiulas. 2019. Model Pruning Enables Efficient Federated Learning on Edge Devices. CoRR, Vol. abs/1909.12326 (2019). [arXiv]1909.12326 http://arxiv.org/abs/1909.12326
[6]
Wenxiang Jiao, Wenxuan Wang, Jen-tse Huang, Xing Wang, and Zhaopeng Tu. 2023. Is ChatGPT A Good Translator? A Preliminary Study. CoRR, Vol. abs/2301.08745 (2023). https://doi.org/10.48550/arXiv.2301.08745 [arXiv]2301.08745
[7]
Armand Joulin, Edouard Grave, Piotr Bojanowski, and Tomá s Mikolov. 2017. Bag of Tricks for Efficient Text Classification. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017, Valencia, Spain, April 3-7, 2017, Volume 2: Short Papers, Mirella Lapata, Phil Blunsom, and Alexander Koller (Eds.). Association for Computational Linguistics, 427--431. https://doi.org/10.18653/v1/e17--2068
[8]
Sai Praneeth Karimireddy, Martin Jaggi, Satyen Kale, Mehryar Mohri, Sashank J. Reddi, Sebastian U. Stich, and Ananda Theertha Suresh. 2021. Breaking the centralized barrier for cross-device federated learning. In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual, Marc'Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (Eds.). 28663--28676. https://proceedings.neurips.cc/paper/2021/hash/f0e6be4ce76ccfa73c5a540d992d0756-Abstract.html
[9]
Anastasia Koloskova, Tao Lin, Sebastian U. Stich, and Martin Jaggi. 2020. Decentralized Deep Learning with Arbitrary Communication Compression. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net. https://openreview.net/forum?id=SkgGCkrKvH
[10]
Yujun Lin, Song Han, Huizi Mao, Yu Wang, and William J. Dally. 2017. Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training. CoRR, Vol. abs/1712.01887 (2017). showeprint[arXiv]1712.01887 http://arxiv.org/abs/1712.01887
[11]
Fan Liu, Hao Liu, and Wenzhao Jiang. 2022. Practical Adversarial Attacks on Spatiotemporal Traffic Forecasting Models. In Advances in Neural Information Processing Systems, Vol. 35. 19035--19047.
[12]
Hao Liu, Ying Li, Yanjie Fu, Huaibo Mei, Jingbo Zhou, Xu Ma, and Hui Xiong. 2020. Polestar: An intelligent, efficient and national-wide public transportation routing engine. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 2321--2329.
[13]
Hao Liu, Qiyu Wu, Fuzhen Zhuang, Xinjiang Lu, Dejing Dou, and Hui Xiong. 2021a. Community-Aware Multi-Task Transportation Demand Prediction. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence. 320--327.
[14]
Jianchun Liu, Hongli Xu, Yang Xu, Zhenguo Ma, Zhiyuan Wang, Chen Qian, and He Huang. 2021b. Communication-efficient asynchronous federated learning in resource-constrained edge computing. Comput. Networks, Vol. 199 (2021), 108429. https://doi.org/10.1016/j.comnet.2021.108429
[15]
Xiaolong Ma, Minghai Qin, Fei Sun, Zejiang Hou, Kun Yuan, Yi Xu, Yanzhi Wang, Yen-Kuang Chen, Rong Jin, and Yuan Xie. 2022. Effective Model Sparsification by Scheduled Grow-and-Prune Methods. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net. https://openreview.net/forum?id=xa6otUDdP2W
[16]
Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Agü era y Arcas. 2017. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, AISTATS 2017, 20-22 April 2017, Fort Lauderdale, FL, USA (Proceedings of Machine Learning Research, Vol. 54), Aarti Singh and Xiaojin (Jerry) Zhu (Eds.). PMLR, 1273--1282. http://proceedings.mlr.press/v54/mcmahan17a.html
[17]
Felix Sattler, Simon Wiedemann, Klaus-Robert Müller, and Wojciech Samek. 2019. Robust and Communication-Efficient Federated Learning from Non-IID Data. CoRR, Vol. abs/1903.02891 (2019). [arXiv]1903.02891 http://arxiv.org/abs/1903.02891
[18]
Navjot Singh, Deepesh Data, Jemin George, and Suhas N. Diggavi. 2021. SQuARM-SGD: Communication-Efficient Momentum SGD for Decentralized Optimization. IEEE J. Sel. Areas Inf. Theory, Vol. 2, 3 (2021), 954--969. https://doi.org/10.1109/JSAIT.2021.3103920
[19]
Hanlin Tang, Xiangru Lian, Shuang Qiu, Lei Yuan, Ce Zhang, Tong Zhang, and Ji Liu. 2019. DeepSqueeze: Parallel Stochastic Gradient Descent with Double-Pass Error-Compensated Compression. CoRR, Vol. abs/1907.07346 (2019). [arXiv]1907.07346 http://arxiv.org/abs/1907.07346
[20]
Chandra Thapa, Mahawaga Arachchige Pathum Chamikara, Seyit Camtepe, and Lichao Sun. 2022. SplitFed: When Federated Learning Meets Split Learning. In Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022. AAAI Press, 8485--8493. https://ojs.aaai.org/index.php/AAAI/article/view/20825
[21]
Zhiyuan Wang, Hongli Xu, Jianchun Liu, He Huang, Chunming Qiao, and Yangming Zhao. 2021. Resource-Efficient Federated Learning with Hierarchical Aggregation in Edge Computing. In 40th IEEE Conference on Computer Communications, INFOCOM 2021, Vancouver, BC, Canada, May 10-13, 2021. IEEE, 1--10. https://doi.org/10.1109/INFOCOM42981.2021.9488756
[22]
Qi Xia, Winson Ye, Zeyi Tao, Jindi Wu, and Qun Li. 2021. A survey of federated learning for edge computing: Research problems and solutions. High-Confidence Computing, Vol. 1, 1 (2021), 100008. https://doi.org/10.1016/j.hcc.2021.100008
[23]
Xin Yao, Tianchi Huang, Rui-Xiao Zhang, Ruiyu Li, and Lifeng Sun. 2019. Federated Learning with Unbiased Gradient Aggregation and Controllable Meta Updating. CoRR, Vol. abs/1910.08234 (2019). showeprint[arXiv]1910.08234 http://arxiv.org/abs/1910.08234
[24]
Hongzheng Yu, Zekai Chen, Xiao Zhang, Xu Chen, Fuzhen Zhuang, Hui Xiong, and Xiuzhen Cheng. 2023. FedHAR: Semi-Supervised Online Learning for Personalized Federated Human Activity Recognition. IEEE Transactions on Mobile Computing, Vol. 22, 6 (2023), 3318--3332. https://doi.org/10.1109/TMC.2021.3136853
[25]
Hao Yu, Sen Yang, and Shenghuo Zhu. 2019. Parallel Restarted SGD with Faster Convergence and Less Communication: Demystifying Why Model Averaging Works for Deep Learning. In The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019. AAAI Press, 5693--5700. https://doi.org/10.1609/aaai.v33i01.33015693
[26]
Binhang Yuan, Cameron R. Wolfe, Chen Dun, Yuxin Tang, Anastasios Kyrillidis, and Chris Jermaine. 2022. Distributed Learning of Fully Connected Neural Networks using Independent Subnet Training. Proc. VLDB Endow., Vol. 15, 8 (2022), 1581--1590. https://www.vldb.org/pvldb/vol15/p1581-wolfe.pdf
[27]
Xiao Zhang, Qilin Wang, Ziming Ye, Haochao Ying, and Dongxiao Yu. 2023 b. Federated Representation Learning With Data Heterogeneity for Human Mobility Prediction. IEEE Transactions on Intelligent Transportation Systems (2023), 1--12. https://doi.org/10.1109/TITS.2023.3252029
[28]
Xiao Zhang, Yangyang Wang, Shuzhen Chen, Cui Wang, Dongxiao Yu, and Xiuzhen Cheng. 2023 a. Robust communication-efficient decentralized learning with heterogeneity. Journal of Systems Architecture (2023), 102900. https://doi.org/10.1016/j.sysarc.2023.102900
[29]
Xiao Zhang, Ziming Ye, Jianfeng Lu, Fuzhen Zhuang, Yanwei Zheng, and Dongxiao Yu. [n.,d.]. Fine-Grained Preference-Aware Personalized Federated POI Recommendation with Data Sparsity. In the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR '23).
[30]
Xiang Zhang, Junbo Jake Zhao, and Yann LeCun. 2015. Character-level Convolutional Networks for Text Classification. In Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, December 7-12, 2015, Montreal, Quebec, Canada, Corinna Cortes, Neil D. Lawrence, Daniel D. Lee, Masashi Sugiyama, and Roman Garnett (Eds.). 649--657. https://proceedings.neurips.cc/paper/2015/hash/250cf8b51c773f3f8dc8b4be867a9a02-Abstract.html
[31]
Hanhan Zhou, Tian Lan, Guru Prasadh Venkataramani, and Wenbo Ding. [n.,d.]. Federated Learning with Online Adaptive Heterogeneous Local Models. In Workshop on Federated Learning: Recent Advances and New Challenges (in Conjunction with NeurIPS 2022).

Cited By

View all
  • (2024)Distributed Learning for Large-Scale Models at Edge With Privacy ProtectionIEEE Transactions on Computers10.1109/TC.2024.335281473:4(1060-1070)Online publication date: 18-Jan-2024
  • (2024)Enhancing neuro-oncology care through equity-driven applications of artificial intelligenceNeuro-Oncology10.1093/neuonc/noae127Online publication date: 19-Aug-2024
  • (2024)BR-FEEL: A backdoor resilient approach for federated edge learning with fragment-sharingJournal of Systems Architecture10.1016/j.sysarc.2024.103258155(103258)Online publication date: Oct-2024
  • Show More Cited By

Index Terms

  1. Theoretical Convergence Guaranteed Resource-Adaptive Federated Learning with Mixed Heterogeneity

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      KDD '23: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
      August 2023
      5996 pages
      ISBN:9798400701030
      DOI:10.1145/3580305
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 04 August 2023

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. federated learning
      2. heterogeneity
      3. limited resources

      Qualifiers

      • Research-article

      Funding Sources

      Conference

      KDD '23
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)585
      • Downloads (Last 6 weeks)30
      Reflects downloads up to 24 Sep 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Distributed Learning for Large-Scale Models at Edge With Privacy ProtectionIEEE Transactions on Computers10.1109/TC.2024.335281473:4(1060-1070)Online publication date: 18-Jan-2024
      • (2024)Enhancing neuro-oncology care through equity-driven applications of artificial intelligenceNeuro-Oncology10.1093/neuonc/noae127Online publication date: 19-Aug-2024
      • (2024)BR-FEEL: A backdoor resilient approach for federated edge learning with fragment-sharingJournal of Systems Architecture10.1016/j.sysarc.2024.103258155(103258)Online publication date: Oct-2024
      • (2024)FedQL: Q-Learning Guided Aggregation for Federated LearningAlgorithms and Architectures for Parallel Processing10.1007/978-981-97-0834-5_16(263-282)Online publication date: 12-Mar-2024
      • (2023)Data Quality Aware Hierarchical Federated Reinforcement Learning Framework for Dynamic Treatment Regimes2023 IEEE International Conference on Data Mining (ICDM)10.1109/ICDM58522.2023.00131(1103-1108)Online publication date: 1-Dec-2023

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Get Access

      Login options

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media