Timezone: »
Asynchronous-parallel algorithms have the potential to vastly speed up algorithms by eliminating costly synchronization. However, our understanding of these algorithms is limited because the current convergence theory of asynchronous block coordinate descent algorithms is based on somewhat unrealistic assumptions. In particular, the age of the shared optimization variables being used to update blocks is assumed to be independent of the block being updated. Additionally, it is assumed that the updates are applied to randomly chosen blocks. In this paper, we argue that these assumptions either fail to hold or will imply less efficient implementations. We then prove the convergence of asynchronous-parallel block coordinate descent under more realistic assumptions, in particular, always without the independence assumption. The analysis permits both the deterministic (essentially) cyclic and random rules for block choices. Because a bound on the asynchronous delays may or may not be available, we establish convergence for both bounded delays and unbounded delays. The analysis also covers nonconvex, weakly convex, and strongly convex functions. The convergence theory involves a Lyapunov function that directly incorporates both objective progress and delays. A continuous-time ODE is provided to motivate the construction at a high level.
Author Information
Tao Sun (National university of defense technology)
College of Science, National University of Defense Technology, PRC.
Robert Hannah (UCLA)
Wotao Yin (University of California, Los Angeles)
More from the Same Authors
-
2022 Poster: Finite-Time Analysis of Adaptive Temporal Difference Learning with Deep Neural Networks »
Tao Sun · Dongsheng Li · Bao Wang -
2019 Poster: General Proximal Incremental Aggregated Gradient Algorithms: Better and Novel Results under General Scheme »
Tao Sun · Yuejiao Sun · Dongsheng Li · Qing Liao -
2018 Poster: LAG: Lazily Aggregated Gradient for Communication-Efficient Distributed Learning »
Tianyi Chen · Georgios Giannakis · Tao Sun · Wotao Yin -
2018 Spotlight: LAG: Lazily Aggregated Gradient for Communication-Efficient Distributed Learning »
Tianyi Chen · Georgios Giannakis · Tao Sun · Wotao Yin -
2018 Poster: Breaking the Span Assumption Yields Fast Finite-Sum Minimization »
Robert Hannah · Yanli Liu · Daniel O'Connor · Wotao Yin -
2018 Poster: On Markov Chain Gradient Descent »
Tao Sun · Yuejiao Sun · Wotao Yin -
2018 Poster: Theoretical Linear Convergence of Unfolded ISTA and Its Practical Weights and Thresholds »
Xiaohan Chen · Jialin Liu · Zhangyang Wang · Wotao Yin -
2018 Spotlight: Theoretical Linear Convergence of Unfolded ISTA and Its Practical Weights and Thresholds »
Xiaohan Chen · Jialin Liu · Zhangyang Wang · Wotao Yin -
2017 Poster: Straggler Mitigation in Distributed Optimization Through Data Encoding »
Can Karakus · Yifan Sun · Suhas Diggavi · Wotao Yin -
2017 Spotlight: Straggler Mitigation in Distributed Optimization Through Data Encoding »
Can Karakus · Yifan Sun · Suhas Diggavi · Wotao Yin