TY - JOUR
T1 - Client Selection for Generalization in Accelerated Federated Learning
T2 - A Multi-Armed Bandit Approach
AU - Ami, Dan Ben
AU - Cohen, Kobi
AU - Zhao, Qing
N1 - Publisher Copyright:
© 2025 Institute of Electrical and Electronics Engineers Inc.. All rights reserved.
PY - 2025/1/1
Y1 - 2025/1/1
N2 - Federated learning (FL) is an emerging machine learning (ML) paradigm used to train models across multiple nodes (i.e., clients) holding local data sets, without explicitly exchanging the data. It has attracted a growing interest in recent years due to its advantages in terms of privacy considerations, and communication resources. In FL, selected clients train their local models and send a function of the models to the server, which consumes a random processing and transmission time. The server updates the global model and broadcasts it back to the clients. The client selection problem in FL is to schedule a subset of the clients for training and transmission at each given time so as to optimize the learning performance. In this paper, we present a novel multi-armed bandit (MAB)-based approach for client selection to minimize the training latency without harming the ability of the model to generalize, that is, to provide reliable predictions for new observations. We develop a novel algorithm to achieve this goal, dubbed Bandit Scheduling for FL (BSFL). We analyze BSFL theoretically, and show that it achieves a logarithmic regret, defined as the loss of BSFL as compared to a genie that has complete knowledge about the latency means of all clients. We conducted evaluations under both i.i.d. and non-i.i.d. scenarios using a synthetic dataset with a linear regression model and two well-known datasets, Fashion-MNIST and CIFAR-10 with CNN-based classification models. The results demonstrate that BSFL outperforms existing methods.
AB - Federated learning (FL) is an emerging machine learning (ML) paradigm used to train models across multiple nodes (i.e., clients) holding local data sets, without explicitly exchanging the data. It has attracted a growing interest in recent years due to its advantages in terms of privacy considerations, and communication resources. In FL, selected clients train their local models and send a function of the models to the server, which consumes a random processing and transmission time. The server updates the global model and broadcasts it back to the clients. The client selection problem in FL is to schedule a subset of the clients for training and transmission at each given time so as to optimize the learning performance. In this paper, we present a novel multi-armed bandit (MAB)-based approach for client selection to minimize the training latency without harming the ability of the model to generalize, that is, to provide reliable predictions for new observations. We develop a novel algorithm to achieve this goal, dubbed Bandit Scheduling for FL (BSFL). We analyze BSFL theoretically, and show that it achieves a logarithmic regret, defined as the loss of BSFL as compared to a genie that has complete knowledge about the latency means of all clients. We conducted evaluations under both i.i.d. and non-i.i.d. scenarios using a synthetic dataset with a linear regression model and two well-known datasets, Fashion-MNIST and CIFAR-10 with CNN-based classification models. The results demonstrate that BSFL outperforms existing methods.
KW - Federated learning (FL)
KW - client scheduling
KW - client selection
KW - generalization in machine learning
KW - multi-armed bandit (MAB)
UR - http://www.scopus.com/inward/record.url?scp=85218742953&partnerID=8YFLogxK
U2 - 10.1109/ACCESS.2025.3543441
DO - 10.1109/ACCESS.2025.3543441
M3 - Article
AN - SCOPUS:85218742953
SN - 2169-3536
JO - IEEE Access
JF - IEEE Access
ER -