Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3369583.3392686acmconferencesArticle/Chapter ViewAbstractPublication PageshpdcConference Proceedingsconference-collections
research-article

TiFL: A Tier-based Federated Learning System

Published: 23 June 2020 Publication History

Abstract

Federated Learning (FL) enables learning a shared model acrossmany clients without violating the privacy requirements. One of the key attributes in FL is the heterogeneity that exists in both resource and data due to the differences in computation and communication capacity, as well as the quantity and content of data among different clients. We conduct a case study to show that heterogeneity in resource and data has a significant impact on training time and model accuracy in conventional FL systems. To this end, we propose TiFL, a Tier-based Federated Learning System, which divides clients into tiers based on their training performance and selects clients from the same tier in each training round to mitigate the straggler problem caused by heterogeneity in resource anddata quantity. To further tame the heterogeneity caused by non-IID (Independent and Identical Distribution) data and resources, TiFL employs an adaptive tier selection approach to update the tiering on-the-fly based on the observed training performance and accuracy. We prototype TiFL in a FL testbed following Google's FL architecture and evaluate it using the state-of-the-art FL benchmarks. Experimental evaluation shows that TiFL outperforms the conventional FL in various heterogeneous conditions. With the proposed adaptive tier selection policy, we demonstrate that TiFL achieves much faster training performance while achieving the same or better test accuracy across the board.

Supplementary Material

MP4 File (3369583.3392686.mp4)
In this presentation, we present our work which demonstrates the effects that data and resource heterogeneity have on Federated learning systems, and based on those observations we propose a system that can reduce the training time of the federated learning process without compromising model accuracy by using an adaptive scheduling policy. This work was done in collaboration with George-Mason University, University of Nevada, Reno, and IBM Research Almaden.

References

[1]
Martín Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, et al. 2016. Tensorflow: A system for large-scale machine learning. In 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16). 265--283.
[2]
Martin Abadi, Andy Chu, Ian Goodfellow, H Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang. 2016. Deep learning with differential privacy. In Proceedings of the 2016 ACMSIGSAC Conference on Computer and Communications Security. ACM, 308--318.
[3]
Amos Beimel, Shiva Prasad Kasiviswanathan, and Kobbi Nissim. 2010. Bounds on the sample complexity for private learning and private data release. In Theory of Cryptography Conference. Springer, 437--454.
[4]
Keith Bonawitz, Hubert Eichner, Wolfgang Grieskamp, Dzmitry Huba, Alex Ingerman, Vladimir Ivanov, Chloe Kiddon, Jakub Konecny, Stefano Mazzocchi, H Brendan McMahan, et al. 2019. Towards Federated Learning at Scale: System Design. arXiv preprint arXiv:1902.01046 (2019).
[5]
Keith Bonawitz, Vladimir Ivanov, Ben Kreuter, Antonio Marcedone, H Brendan McMahan, Sarvar Patel, Daniel Ramage, Aaron Segal, and Karn Seth. 2017. Practical secure aggregation for privacy-preserving machine learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security. ACM, 1175--1191.
[6]
Sebastian Caldas, Jakub Konecny, H Brendan McMahan, and Ameet Talwalkar. 2018. Expanding the Reach of Federated Learning by Reducing Client Resource Requirements. arXiv preprint arXiv:1812.07210 (2018).
[7]
Sebastian Caldas, Peter Wu, Tian Li, Jakub Konecny, H Brendan McMahan, Virginia Smith, and Ameet Talwalkar. 2018. Leaf: A benchmark for federated settings. arXiv preprint arXiv:1812.01097 (2018).
[8]
Zheng Chai, Hannan Fayyaz, Zeshan Fayyaz, Ali Anwar, Yi Zhou, Nathalie Baracaldo, Heiko Ludwig, and Yue Cheng. 2019. Towards Taming the Resource and Data Heterogeneity in Federated Learning. In 2019 USENIX Conference on Operational Machine Learning (OpML 19). 19--21.
[9]
Avishek Ghosh, Justin Hong, Dong Yin, and Kannan Ramchandran. 2019. Robust Federated Learning in a Heterogeneous Environment. arXiv preprint arXiv:1906.06629 (2019).
[10]
Aaron Harlap, Henggang Cui,Wei Dai, JinliangWei, Gregory R Ganger, Phillip B Gibbons, Garth A Gibson, and Eric P Xing. 2016. Addressing the straggler problem for iterative convergent parallel ML. In Proceedings of the Seventh ACM Symposium on Cloud Computing. ACM, 98--111.
[11]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. 770--778.
[12]
https://github.com/PaddlePaddle/PaddleFL 2020. PaddleFL. [Online; accessed 21-January-2020].
[13]
https://www.tensorflow.org/federated 2020. Tensorflow Federated. [Online; accessed 21-January-2020].
[14]
Peter Kairouz, H Brendan McMahan, Brendan Avent, Aurélien Bellet, Mehdi Bennis, Arjun Nitin Bhagoji, Keith Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, et al. 2019. Advances and open problems in federated learning. arXiv preprint arXiv:1912.04977 (2019).
[15]
Jakub Konecny, H Brendan McMahan, Felix X Yu, Peter Richtárik, Ananda Theertha Suresh, and Dave Bacon. 2016. Federated learning: Strategies for improving communication efficiency. arXiv preprint arXiv:1610.05492 (2016).
[16]
Alex Krizhevsky, Vinod Nair, and Geoffrey Hinton. 2014. The CIFAR-10 dataset. online: http://www.cs.toronto.edu/kriz/cifar.html 55 (2014).
[17]
Tian Li, Anit Kumar Sahu, Ameet Talwalkar, and Virginia Smith. 2019. Federated learning: Challenges, methods, and future directions. arXiv preprint arXiv:1908.07873 (2019).
[18]
Tian Li, Anit Kumar Sahu, Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, and Virginia Smith. 2018. Federated optimization in heterogeneous networks. arXiv preprint arXiv:1812.06127 (2018).
[19]
Lumin Liu, Jun Zhang, SH Song, and Khaled B Letaief. 2019. Edge-Assisted Hierarchical Federated Learning with Non-IID Data. arXiv preprint arXiv:1905.06641 (2019).
[20]
H Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, et al. 2016. Communication-efficient learning of deep networks from decentralized data. arXiv preprint arXiv:1602.05629 (2016).
[21]
HBrendan McMahan, Daniel Ramage, Kunal Talwar, and Li Zhang. 2017. Learning differentially private recurrent language models. arXiv preprint arXiv:1710.06963 (2017).
[22]
Milad Nasr, Reza Shokri, and Amir Houmansadr. 2018. Comprehensive privacy analysis of deep learning: Stand-alone and federated learning under passive and active white-box inference attacks. arXiv preprint arXiv:1812.00910 (2018).
[23]
Takayuki Nishio and Ryo Yonetani. 2019. Client selection for federated learning with heterogeneous resources in mobile edge. In ICC 2019--2019 IEEE International Conference on Communications (ICC). IEEE, 1--7.
[24]
Jacquelyn K O'herrin, Norman Fost, and Kenneth AKudsk. 2004. Health Insurance Portability Accountability Act (HIPAA) regulations: effect on medical record research. Annals of surgery 239, 6 (2004), 772.
[25]
Theo Ryffel, Andrew Trask, Morten Dahl, Bobby Wagner, Jason Mancuso, Daniel Rueckert, and Jonathan Passerat-Palmbach. 2018. A generic framework for privacy preserving deep learning. arXiv preprint arXiv:1811.04017 (2018).
[26]
Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. 2017. Membership inference attacks against machine learning models. In 2017 IEEE Symposium on Security and Privacy (SP). IEEE, 3--18.
[27]
Colin Tankard. 2016. What the GDPR means for businesses. Network Security 2016, 6 (2016), 5--8.
[28]
Han Xiao, Kashif Rasul, and Roland Vollgraf. 2017. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747 (2017).
[29]
Qiang Yang, Yang Liu, Tianjian Chen, and Yongxin Tong. 2019. Federated machine learning: Concept and applications. ACM Transactions on Intelligent Systems and Technology (TIST) 10, 2 (2019), 12.
[30]
Chengliang Zhang, Huangshi Tian, Wei Wang, and Feng Yan. 2018. Stay Fresh: Speculative Synchronization for Fast Distributed Machine Learning. In 2018 IEEE 38th International Conference on Distributed Computing Systems (ICDCS). IEEE, 99--109.
[31]
Yue Zhao, Meng Li, Liangzhen Lai, Naveen Suda, Damon Civin, and Vikas Chandra. 2018. Federated learning with non-iid data. arXiv preprint arXiv:1806.00582 (2018).

Cited By

View all
  • (2025)FedCust: Offloading hyperparameter customization for federated learningPerformance Evaluation10.1016/j.peva.2024.102450167(102450)Online publication date: Mar-2025
  • (2024)An Efficient Attribute-Based Participant Selecting Scheme with Blockchain for Federated Learning in Smart CitiesComputers10.3390/computers1305011813:5(118)Online publication date: 9-May-2024
  • (2024)QuoTa: An Online Quality-Aware Incentive Mechanism for Fast Federated LearningApplied Sciences10.3390/app1402083314:2(833)Online publication date: 18-Jan-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
HPDC '20: Proceedings of the 29th International Symposium on High-Performance Parallel and Distributed Computing
June 2020
246 pages
ISBN:9781450370523
DOI:10.1145/3369583
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 23 June 2020

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. data heterogeneity
  2. edge computing
  3. federated learning
  4. non-IID
  5. resource heterogeneity
  6. stragglers

Qualifiers

  • Research-article

Funding Sources

Conference

HPDC '20
Sponsor:

Acceptance Rates

Overall Acceptance Rate 166 of 966 submissions, 17%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)472
  • Downloads (Last 6 weeks)36
Reflects downloads up to 18 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2025)FedCust: Offloading hyperparameter customization for federated learningPerformance Evaluation10.1016/j.peva.2024.102450167(102450)Online publication date: Mar-2025
  • (2024)An Efficient Attribute-Based Participant Selecting Scheme with Blockchain for Federated Learning in Smart CitiesComputers10.3390/computers1305011813:5(118)Online publication date: 9-May-2024
  • (2024)QuoTa: An Online Quality-Aware Incentive Mechanism for Fast Federated LearningApplied Sciences10.3390/app1402083314:2(833)Online publication date: 18-Jan-2024
  • (2024)FedNIC: enhancing privacy-preserving federated learning via homomorphic encryption offload on SmartNICFrontiers in Computer Science10.3389/fcomp.2024.14653526Online publication date: 21-Oct-2024
  • (2024)FedCaSe: Enhancing Federated Learning with Heterogeneity-aware Caching and SchedulingProceedings of the 2024 ACM Symposium on Cloud Computing10.1145/3698038.3698559(52-68)Online publication date: 20-Nov-2024
  • (2024)Topology-aware Federated Learning in Edge Computing: A Comprehensive SurveyACM Computing Surveys10.1145/365920556:10(1-41)Online publication date: 22-Jun-2024
  • (2024)Ensuring Fairness and Gradient Privacy in Personalized Heterogeneous Federated LearningACM Transactions on Intelligent Systems and Technology10.1145/365261315:3(1-30)Online publication date: 13-Mar-2024
  • (2024)Privacy-Preserving and Cross-Domain Human Sensing by Federated Domain Adaptation with Semantic Knowledge CorrectionProceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/36435038:1(1-26)Online publication date: 6-Mar-2024
  • (2024)Behave Differently when Clustering: A Semi-asynchronous Federated Learning Approach for IoTACM Transactions on Sensor Networks10.1145/363982520:3(1-28)Online publication date: 25-Jan-2024
  • (2024)Totoro: A Scalable Federated Learning Engine for the EdgeProceedings of the Nineteenth European Conference on Computer Systems10.1145/3627703.3629575(182-199)Online publication date: 22-Apr-2024
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media