TY - JOUR
T1 - Fast-Convergent Federated Learning
AU - Nguyen, Hung T.
AU - Sehwag, Vikash
AU - Hosseinalipour, Seyyedali
AU - Brinton, Christopher G.
AU - Chiang, Mung
AU - Vincent Poor, H.
N1 - Funding Information:
Manuscript received July 27, 2020; revised September 27, 2020; accepted October 21, 2020. Date of publication November 9, 2020; date of current version December 16, 2020. The work of Hung T. Nguyen and Mung Chiang was supported in part by the Defense Advanced Research Projects Agency (DARPA) under Contract AWD1005371 and Contract AWD1005468. The work of H. Vincent Poor was supported in part by the U.S. National Science Foundation under Grant CCF-1908308. (Corresponding author: Hung T. Nguyen.) Hung T. Nguyen, Vikash Sehwag, and H. Vincent Poor are with the Department of Electrical Engineering, Princeton University, Princeton, NJ 08544 USA (e-mail: hn4@princeton.edu; vvikash@princeton.edu; poor@princeton.edu).
Publisher Copyright:
© 1983-2012 IEEE.
PY - 2021/1
Y1 - 2021/1
N2 - Federated learning has emerged recently as a promising solution for distributing machine learning tasks through modern networks of mobile devices. Recent studies have obtained lower bounds on the expected decrease in model loss that is achieved through each round of federated learning. However, convergence generally requires a large number of communication rounds, which induces delay in model training and is costly in terms of network resources. In this paper, we propose a fast-convergent federated learning algorithm, called $\mathsf {FOLB}$ , which performs intelligent sampling of devices in each round of model training to optimize the expected convergence speed. We first theoretically characterize a lower bound on improvement that can be obtained in each round if devices are selected according to the expected improvement their local models will provide to the current global model. Then, we show that $\mathsf {FOLB}$ obtains this bound through uniform sampling by weighting device updates according to their gradient information. $\mathsf {FOLB}$ is able to handle both communication and computation heterogeneity of devices by adapting the aggregations according to estimates of device's capabilities of contributing to the updates. We evaluate $\mathsf {FOLB}$ in comparison with existing federated learning algorithms and experimentally show its improvement in trained model accuracy, convergence speed, and/or model stability across various machine learning tasks and datasets.
AB - Federated learning has emerged recently as a promising solution for distributing machine learning tasks through modern networks of mobile devices. Recent studies have obtained lower bounds on the expected decrease in model loss that is achieved through each round of federated learning. However, convergence generally requires a large number of communication rounds, which induces delay in model training and is costly in terms of network resources. In this paper, we propose a fast-convergent federated learning algorithm, called $\mathsf {FOLB}$ , which performs intelligent sampling of devices in each round of model training to optimize the expected convergence speed. We first theoretically characterize a lower bound on improvement that can be obtained in each round if devices are selected according to the expected improvement their local models will provide to the current global model. Then, we show that $\mathsf {FOLB}$ obtains this bound through uniform sampling by weighting device updates according to their gradient information. $\mathsf {FOLB}$ is able to handle both communication and computation heterogeneity of devices by adapting the aggregations according to estimates of device's capabilities of contributing to the updates. We evaluate $\mathsf {FOLB}$ in comparison with existing federated learning algorithms and experimentally show its improvement in trained model accuracy, convergence speed, and/or model stability across various machine learning tasks and datasets.
KW - Federated learning
KW - distributed optimization
KW - fast convergence rate
UR - http://www.scopus.com/inward/record.url?scp=85096379296&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85096379296&partnerID=8YFLogxK
U2 - 10.1109/JSAC.2020.3036952
DO - 10.1109/JSAC.2020.3036952
M3 - Article
AN - SCOPUS:85096379296
SN - 0733-8716
VL - 39
SP - 201
EP - 218
JO - IEEE Journal on Selected Areas in Communications
JF - IEEE Journal on Selected Areas in Communications
IS - 1
M1 - 9252927
ER -