ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.04374
15
16

Fast and Robust Distributed Learning in High Dimension

5 May 2019
El-Mahdi El-Mhamdi
R. Guerraoui
Sébastien Rouault
    OOD
    FedML
ArXivPDFHTML
Abstract

Could a gradient aggregation rule (GAR) for distributed machine learning be both robust and fast? This paper answers by the affirmative through multi-Bulyan. Given nnn workers, fff of which are arbitrary malicious (Byzantine) and m=n−fm=n-fm=n−f are not, we prove that multi-Bulyan can ensure a strong form of Byzantine resilience, as well as an mn{\frac{m}{n}}nm​ slowdown, compared to averaging, the fastest (but non Byzantine resilient) rule for distributed machine learning. When m≈nm \approx nm≈n (almost all workers are correct), multi-Bulyan reaches the speed of averaging. We also prove that multi-Bulyan's cost in local computation is O(d)O(d)O(d) (like averaging), an important feature for ML where ddd commonly reaches 10910^9109, while robust alternatives have at least quadratic cost in ddd. Our theoretical findings are complemented with an experimental evaluation which, in addition to supporting the linear O(d)O(d)O(d) complexity argument, conveys the fact that multi-Bulyan's parallelisability further adds to its efficiency.

View on arXiv
Comments on this paper