Abstract
Federated learning is a distributed machine learning paradigm that allows multiple edge devices to collaboratively train a shared model without exchanging raw data. However, the training efficiency of federated learning is highly dependent on client selection. Moreover, due to the varying wireless communication environments and various computation latencies among the clients, selecting clients randomly or uniformly may not be optimal for balancing the data diversity and training efficiency. In this article, we formulate a new latency-minimization problem that simultaneously optimizes client selection and training procedures in federated learning, which takes into account the data and latency heterogeneity among the clients. Given the nonconvexity of the problem, we derive a new convergence upper bound for federated learning with probabilistic client selection. To solve the mixed integer nonlinear programming problem, we introduce a hybrid solution that integrates grid search techniques with the polyhedral active set algorithm. Numerical analyses and experiments on real-world data demonstrate that our scheme outperforms the existing ones in terms of overall training latency and achieves up to three times acceleration over random client selection, especially in scenarios with highly heterogeneous data and latencies among the clients.
Original language | English (US) |
---|---|
Pages (from-to) | 32183-32196 |
Number of pages | 14 |
Journal | IEEE Internet of Things Journal |
Volume | 11 |
Issue number | 19 |
DOIs | |
State | Published - 2024 |
Externally published | Yes |
All Science Journal Classification (ASJC) codes
- Signal Processing
- Information Systems
- Hardware and Architecture
- Computer Science Applications
- Computer Networks and Communications
Keywords
- Client selection
- data heterogeneity
- federated learning
- latency heterogeneity
- optimization