Hongbin ZhuMiao YangJunqian KuangHua QianYong Zhou
Federated learning (FL), as a nascent distributed learning framework, trains a machine learning model in a collaborative manner. Synchronous model aggregation is widely adopted, but suffers from the straggler issue because of the system heterogeneity. To overcome the straggler issue, we employ the asynchronous FL framework. The target of this paper is to minimize the training latency by client selection while taking into account both the client availability and the long-term fairness. A practical scenario is considered where the channel conditions and the local computing power of the clients are not aware by the parameter server. This makes client selection problem thorny to be tackled, because the training latency consists of time-varying round trip transmission latency and the local training latency. By transforming the latency minimization problem into a multi-armed bandit problem and leveraging the upper confidence bound policy and the virtual queue technique, we tackle the asynchronous client selection problem. Numerical results validate that our proposed algorithm outperforms the baseline algorithms in terms of the convergence performance.
Hongbin ZhuYong ZhouHua QianYuanming ShiXu ChenYang Yang
Jiarong YangYuan LiuFangjiong ChenWen ChenChangle Li
Yuxin ShiZelei LiuZhuan ShiHan Yu
Hongbin ZhuJunqian KuangMiao YangHua Qian
Xuerui LiYangming ZhaoChunming Qiao