TY - GEN
T1 - Submodel Partitioning in Hierarchical Federated Learning
T2 - 59th Annual IEEE International Conference on Communications, ICC 2024
AU - Fang, Wenzhi
AU - Han, Dong Jun
AU - Brinton, Christopher G.
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Hierarchical federated learning (HFL) has demon-strated promising scalability advantages over the traditional 'star-topology' architecture-based federated learning (FL). How-ever, HFL still imposes significant computation, communication, and storage burdens on the edge, especially when training a large-scale model over resource-constrained Internet of Things (IoT) devices. In this paper, we propose hierarchical independent submodel training (HIST), a new FL methodology that aims to address these issues in hierarchical settings. The key idea behind HIST is a hierarchical version of model partitioning, where we partition the global model into disjoint submodels in each round, and distribute them across different cells, so that each cell is responsible for training only one partition of the full model. This enables each client to save computation/storage costs while alleviating the communication loads throughout the hierarchy. We characterize the convergence behavior of HIST for non-convex loss functions under mild assumptions, showing the impact of several attributes (e.g., number of cells, local and global aggregation frequency) on the performance-efficiency tradeoff. Finally, through numerical experiments, we verify that HIST is able to save communication costs by a wide margin while achieving the same target testing accuracy.
AB - Hierarchical federated learning (HFL) has demon-strated promising scalability advantages over the traditional 'star-topology' architecture-based federated learning (FL). How-ever, HFL still imposes significant computation, communication, and storage burdens on the edge, especially when training a large-scale model over resource-constrained Internet of Things (IoT) devices. In this paper, we propose hierarchical independent submodel training (HIST), a new FL methodology that aims to address these issues in hierarchical settings. The key idea behind HIST is a hierarchical version of model partitioning, where we partition the global model into disjoint submodels in each round, and distribute them across different cells, so that each cell is responsible for training only one partition of the full model. This enables each client to save computation/storage costs while alleviating the communication loads throughout the hierarchy. We characterize the convergence behavior of HIST for non-convex loss functions under mild assumptions, showing the impact of several attributes (e.g., number of cells, local and global aggregation frequency) on the performance-efficiency tradeoff. Finally, through numerical experiments, we verify that HIST is able to save communication costs by a wide margin while achieving the same target testing accuracy.
UR - https://www.scopus.com/pages/publications/85202815859
UR - https://www.scopus.com/inward/citedby.url?scp=85202815859&partnerID=8YFLogxK
U2 - 10.1109/ICC51166.2024.10622512
DO - 10.1109/ICC51166.2024.10622512
M3 - Conference contribution
AN - SCOPUS:85202815859
T3 - IEEE International Conference on Communications
SP - 268
EP - 273
BT - ICC 2024 - IEEE International Conference on Communications
A2 - Valenti, Matthew
A2 - Reed, David
A2 - Torres, Melissa
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 9 June 2024 through 13 June 2024
ER -