Distributed Optimization, Averaging via ADMM, and Network Topology

From MaRDI portal
Publication:6348485

arXiv2009.02604MaRDI QIDQ6348485FDOQ6348485


Authors: Guilherme França, José Bento Edit this on Wikidata


Publication date: 5 September 2020

Abstract: There has been an increasing necessity for scalable optimization methods, especially due to the explosion in the size of datasets and model complexity in modern machine learning applications. Scalable solvers often distribute the computation over a network of processing units. For simple algorithms such as gradient descent the dependency of the convergence time with the topology of this network is well-known. However, for more involved algorithms such as the Alternating Direction Methods of Multipliers (ADMM) much less is known. At the heart of many distributed optimization algorithms there exists a gossip subroutine which averages local information over the network, and whose efficiency is crucial for the overall performance of the method. In this paper we review recent research in this area and, with the goal of isolating such a communication exchange behaviour, we compare different algorithms when applied to a canonical distributed averaging consensus problem. We also show interesting connections between ADMM and lifted Markov chains besides providing an explicitly characterization of its convergence and optimal parameter tuning in terms of spectral properties of the network. Finally, we empirically study the connection between network topology and convergence rates for different algorithms on a real world problem of sensor localization.




Has companion code repository: https://github.com/bentoayr/distributed-opt-and-topology









This page was built for publication: Distributed Optimization, Averaging via ADMM, and Network Topology

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6348485)