Bregman parallel direction method of multipliers (BPDMM) efficiently solves distributed optimization over a network, which arises in a wide spectrum of collaborative multi-agent learning applications. In this paper, we generalize BPDMM to stochastic BPDMM, where each iteration only solves local optimization on a randomly selected subset of nodes rather than all the nodes in the network. Such generalization reduce the need for computational resources and allows applications to larger scale networks. We establish both the global convergence and the O(1/T) iteration complexity of stochastic BPDMM. We demonstrate our results via numerical examples.
References
[1]
M. Mesbahi and M. Egerstedt, Graph Theoretic Methods in Multiagent Networks. Princeton University Press, 2010.
D. Li, K. D. Wong, Y. H. Hu, and A. M. Sayeed, “Detection, classification, and tracking of targets,” IEEE Signal Process. Mag., vol. 19, no. 2, pp. 17–29, 2002.
B. Açıkmeşe, M. Mandić, and J. L. Speyer, “Decentralized observers with consensus filters for distributed discrete-time linear systems,” Automatica, vol. 50, no. 4, pp. 1037–1052, 2014.
B. Gholami, S. Yoon, and V. Pavlovic, “Decentralized approximate bayesian inference for distributed sensor network.” in AAAI Conf. Artificial Intell., 2016, pp. 1582–1588.
A. Yahya, A. Li, M. Kalakrishnan, Y. Chebotar, and S. Levine, “Collective robot reinforcement learning with distributed asynchronous guided policy search,” in Int. Conf. Intell. Robots Syst. IEEE, 2017, pp. 79–86.
D. Meng, M. Fazel, and M. Mesbahi, “Proximal alternating direction method of multipliers for distributed optimization on weighted graphs,” in Proc. IEEE Conf. Decision Control, 2015, pp. 1396–1401.
Y. Yu, B. Açıkmes¸e, and M. Mesbahi, “Bregman parallel direction method of multipliers for distributed optimization via mirror averaging,” IEEE Control Syst. Lett., vol. 2, no. 2, pp. 302–306, 2018.
E. Wei and A. Ozdaglar, “On the O(1/k) convergence of asynchronous distributed alternating direction method of multipliers,” in Proc. Global Conf. Signal Inform. Process. IEEE, 2013, pp. 551–554.
Z. Zhu and A. J. Storkey, “Stochastic parallel block coordinate descent for large-scale saddle point problems.” in Proc. AAAI Conf. Artificial Intell., 2016, pp. 2429–2437.
P. Richtárik and M. Takáč, “Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function,” Math. Prog., vol. 144, no. 1-2, pp. 1–38, 2014.
S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, “Distributed optimization and statistical learning via the alternating direction method of multipliers,” Found. Trends Mach. Learn., vol. 3, no. 1, pp. 1–122, 2011.
D. Jakovetic, J. M. Moura, and J. Xavier, “Distributed augmented lagrangian algorithms: convergence rate,” in Proc. IEEE Global Conf. Signal Inform. Process., 2013, pp. 563–566.
F. Iutzeler, P. Bianchi, P. Ciblat, and W. Hachem, “Asynchronous distributed optimization using a randomized alternating direction method of multipliers,” in Proc. IEEE Conf. Decision Control, 2013, pp. 3671–3676.
W. Shi, Q. Ling, K. Yuan, G. Wu, and W. Yin, “On the linear convergence of the ADMM in decentralized consensus optimization.” IEEE Trans. Signal Process., vol. 62, no. 7, pp. 1750–1761, 2014.
Y. Yu, B. Açıkmes¸e, and M. Mesbahi, “Mass-spring-damper network for distributed averaging and optimization,” arXiv preprint arXiv:1808.01999 [math. OC], 2018.
Y. Yu and B. Açıkmeşe, “Stochastic bregman parallel direction method of multipliers for distributed optimization,” arXiv preprint arXiv:1902.09695 [math.OC], 20119.