Linear Convergence of Asynchronous Gradient Push Algorithm for Distributed Optimization

This article focuses on multiagent distributed asynchronous optimization over directed networks where each agent can only access its individual local function, and the aggregate aim is to minimize the cumulative sum of all local functions. Considering the asynchrony among the agents, we develop an a...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on systems, man, and cybernetics. Systems Vol. 55; no. 3; pp. 2147 - 2159
Main Authors: Li, Huaqing, Cheng, Huqiang, Lu, Qingguo, Wang, Zheng, Huang, Tingwen
Format: Journal Article
Language:English
Published: IEEE 01.03.2025
Subjects:
ISSN:2168-2216, 2168-2232
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This article focuses on multiagent distributed asynchronous optimization over directed networks where each agent can only access its individual local function, and the aggregate aim is to minimize the cumulative sum of all local functions. Considering the asynchrony among the agents, we develop an algorithm in which agents compute and communicate individually, without any form of synchronized coordination. Agents perform their local updates by local communication with their immediate neighbors, and this may involve the use of stale information. Since asynchrony naturally leads to latency or packet loss, an asynchronous robust gradient tracking mechanism is developed to guarantee estimating the average of agents' gradients precisely. Moreover, it employs uncoordinated step-sizes which are more flexible and general than constant or decaying step-size. When the global objective is strongly convex and the local objectives have Lipschitz-continuous gradients, we prove that each agent executing the asynchronous algorithm linearly converges to the consensus optimal point at an <inline-formula> <tex-math notation="LaTeX">\mathcal {O}(\lambda ^{k}) </tex-math></inline-formula> rate, where <inline-formula> <tex-math notation="LaTeX">\lambda \in (0,1) </tex-math></inline-formula> is convergence factor and k represents the iteration number, with a step satisfying a tight explicit upper bound. Numerical experiments demonstrate that our algorithm has better advantages over the state-of-the-art asynchronous algorithms.
ISSN:2168-2216
2168-2232
DOI:10.1109/TSMC.2024.3516936