计算机科学
修剪
加速
瓶颈
架空(工程)
分布式计算
计算
同步(交流)
过程(计算)
趋同(经济学)
方案(数学)
残余物
机器学习
人工智能
计算机工程
并行计算
计算机网络
算法
频道(广播)
嵌入式系统
数学分析
数学
农学
经济
生物
经济增长
操作系统
作者
Zhida Jiang,Yang Xu,Hongli Xu,Zhiyuan Wang,Jianchun Liu,Qian Chen,Chunming Qiao
标识
DOI:10.1109/tmc.2023.3247798
摘要
Federated learning (FL) has emerged as a promising distributed learning paradigm that enables a large number of mobile devices to cooperatively train a model without sharing their raw data. The iterative training process of FL incurs considerable computation and communication overhead. The workers participating in FL are usually heterogeneous and the workers with poor capabilities may become the bottleneck of model training. To address the challenges of resource overhead and system heterogeneity, this article proposes an efficient FL framework, called FedMP, that improves both computation and communication efficiency over heterogeneous workers through adaptive model pruning. We theoretically analyze the impact of pruning ratio on training performance, and employ a Multi-Armed Bandit based online learning algorithm to adaptively determine different pruning ratios for heterogeneous workers, even without any prior knowledge of their capabilities. As a result, each worker in FedMP can train and transmit the sub-model that fits its own capabilities, accelerating the training process without hurting model accuracy. To prevent the diverse structures of pruned models from affecting the training convergence, we further present a new parameter synchronization scheme, called Residual Recovery Synchronous Parallel (R2SP). Besides, our proposed framework can be extended to the peer-to-peer (P2P) setting. Extensive experiments on physical devices demonstrate that FedMP is effective for different heterogeneous scenarios and data distributions, and can provide up to 4.1× speedup compared to the existing FL methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI