Accelerating decentralized federated learning via momentum GD with heterogeneous delays
Na Li , Hangguan Shan , Meiyan Song , Yong Zhou , Zhongyuan Zhao , Howard H. Yang , Fen Hou
High-Confidence Computing ›› 2025, Vol. 5 ›› Issue (4) : 100310
Accelerating decentralized federated learning via momentum GD with heterogeneous delays
Federated learning (FL) with synchronous model aggregation suffers from the straggler issue because of heterogeneous transmission and computation delays among different agents. In mobile wireless networks, this issue is exacerbated by time-varying network topology due to agent mobility. Although asynchronous FL can alleviate straggler issues, it still faces critical challenges in terms of algorithm design and convergence analysis because of dynamic information update delay (IU-Delay) and dynamic network topology. To tackle these challenges, we propose a decentralized FL framework based on gradient descent with momentum, named decentralized momentum federated learning (DMFL). We prove that DMFL is globally convergent on convex loss functions under the bounded time-varying IU-Delay, as long as the network topology is uniformly jointly strongly connected. Moreover, DMFL does not impose any restrictions on the data distribution over agents. Extensive experiments are conducted to verify DMFL’s performance superiority over the benchmarks and to reveal the effects of diverse parameters on the performance of the proposed algorithm.
Decentralized federated learning / Gradient descent / Momentum / Information update delay / Convergence
| [1] |
|
| [2] |
Enrique Tomás Martínez Beltrán, |
| [3] |
|
| [4] |
|
| [5] |
|
| [6] |
|
| [7] |
|
| [8] |
|
| [9] |
|
| [10] |
|
| [11] |
|
| [12] |
|
| [13] |
|
| [14] |
|
| [15] |
|
| [16] |
|
| [17] |
|
| [18] |
|
| [19] |
|
| [20] |
|
| [21] |
|
| [22] |
|
| [23] |
|
| [24] |
|
| [25] |
|
| [26] |
|
| [27] |
Durmus Alp Emre Acar, |
| [28] |
|
| [29] |
|
/
| 〈 |
|
〉 |