Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3087801.3087861acmconferencesArticle/Chapter ViewAbstractPublication PagespodcConference Proceedingsconference-collections
abstract

Brief Announcement: Byzantine-Tolerant Machine Learning

Published: 25 July 2017 Publication History

Abstract

We report on Krum, the first provably Byzantine-tolerant aggregation rule for distributed Stochastic Gradient Descent (SGD). Krum guarantees the convergence of SGD even in a distributed setting where (asymptotically) up to half of the workers can be malicious adversaries trying to attack the learning system.

References

[1]
M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving, M. Isard, et al. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI). Savannah, Georgia, USA, 2016.
[2]
P. Blanchard, E.M. El Mhamdi, R. Guerraoui, J. Stainer. Byzantine-TolerantMachine Learning. In arXiv preprint arXiv:1703.02757, 2017.
[3]
L. Bottou. Online learning and stochastic approximations. Online learning in neural networks, 17(9):142, 1998.
[4]
L. Bottou. Large-scale machine learning with stochastic gradient descent. In Proceedings of COMPSTAT'2010, pages 177--186. Springer, 2010.
[5]
L. Lamport, R. Shostak, and M. Pease. The byzantine generals problem. ACM Transactions on Programming Languages and Systems (TOPLAS), 4(3):382--401, 1982.
[6]
X. Lian, Y. Huang, Y. Li, and J. Liu. Asynchronous parallel stochastic gradient for nonconvex optimization. In Advances in Neural Information Processing Systems, pages 2737--2745, 2015.
[7]
N. A. Lynch. Distributed algorithms. Morgan Kaufmann, 1996.
[8]
J. Markoff. How many computers to identify a cat? 16,000. New York Times, pages 06--25, 2012.
[9]
H. Mendes and M. Herlihy. Multidimensional approximate agreement in byzantine asynchronous systems. In Proceedings of the forty-fifth annual ACM symposium on Theory of computing, pages 391--400. ACM, 2013.
[10]
B. T. Polyak and A. B. Juditsky. Acceleration of stochastic approximation by averaging. SIAM Journal on Control and Optimization, 30(4):838--855, 1992.
[11]
F. B. Schneider. Implementing fault-tolerant services using the state machine approach: A tutorial. ACM Computing Surveys (CSUR), 22(4):299--319, 1990.
[12]
S. Zhang, A. E. Choromanska, and Y. LeCun. Deep learning with elastic averaging sgd. In Advances in Neural Information Processing Systems, pages 685--693, 2015.

Cited By

View all
  • (2024)Privacy Preserving Technology using Federated Learning and Blockchain in protecting against Adversarial Attacks for Retinal ImagingOphthalmology10.1016/j.ophtha.2024.10.017Online publication date: Oct-2024
  • (2023)Byzantine-Resilient Learning Beyond Gradients: Distributing Evolutionary SearchProceedings of the Companion Conference on Genetic and Evolutionary Computation10.1145/3583133.3590719(295-298)Online publication date: 15-Jul-2023
  • (2017)Machine learning with adversariesProceedings of the 31st International Conference on Neural Information Processing Systems10.5555/3294771.3294783(118-128)Online publication date: 4-Dec-2017

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
PODC '17: Proceedings of the ACM Symposium on Principles of Distributed Computing
July 2017
480 pages
ISBN:9781450349925
DOI:10.1145/3087801
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 25 July 2017

Check for updates

Author Tags

  1. adversarial machine learning
  2. attacking machine learning
  3. distributed stochastic gradient descent

Qualifiers

  • Abstract

Funding Sources

  • The Swiss National Science Foundation
  • European ERC

Conference

PODC '17
Sponsor:

Acceptance Rates

PODC '17 Paper Acceptance Rate 38 of 154 submissions, 25%;
Overall Acceptance Rate 740 of 2,477 submissions, 30%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)7
  • Downloads (Last 6 weeks)1
Reflects downloads up to 15 Oct 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Privacy Preserving Technology using Federated Learning and Blockchain in protecting against Adversarial Attacks for Retinal ImagingOphthalmology10.1016/j.ophtha.2024.10.017Online publication date: Oct-2024
  • (2023)Byzantine-Resilient Learning Beyond Gradients: Distributing Evolutionary SearchProceedings of the Companion Conference on Genetic and Evolutionary Computation10.1145/3583133.3590719(295-298)Online publication date: 15-Jul-2023
  • (2017)Machine learning with adversariesProceedings of the 31st International Conference on Neural Information Processing Systems10.5555/3294771.3294783(118-128)Online publication date: 4-Dec-2017

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media