计算机科学
水准点(测量)
推荐系统
过程(计算)
事实上
培训(气象学)
集合(抽象数据类型)
联合学习
训练集
机器学习
数据库
人工智能
多媒体
气象学
地理
法学
程序设计语言
大地测量学
物理
操作系统
政治学
作者
Khalil Muhammad,Qinqin Wang,Diarmuid O'Reilly-Morgan,Ηλίας Τράγος,Barry Smyth,Neil Hurley,James R. Geraci,Aonghus Lawlor
出处
期刊:Knowledge Discovery and Data Mining
日期:2020-08-20
被引量:111
标识
DOI:10.1145/3394486.3403176
摘要
Federated learning (FL) is quickly becoming the de facto standard for the distributed training of deep recommendation models, using on-device user data and reducing server costs. In a typical FL process, a central server tasks end-users to train a shared recommendation model using their local data. The local models are trained over several rounds on the users' devices and the server combines them into a global model, which is sent to the devices for the purpose of providing recommendations. Standard FL approaches use randomly selected users for training at each round, and simply average their local models to compute the global model. The resulting federated recommendation models require significant client effort to train and many communication rounds before they converge to a satisfactory accuracy. Users are left with poor quality recommendations until the late stages of training. We present a novel technique, FedFast, to accelerate distributed learning which achieves good accuracy for all users very early in the training process. We achieve this by sampling from a diverse set of participating clients in each training round and applying an active aggregation method that propagates the updated model to the other clients. Consequently, with FedFast the users benefit from far lower communication costs and more accurate models that can be consumed anytime during the training process even at the very early stages. We demonstrate the efficacy of our approach across a variety of benchmark datasets and in comparison to state-of-the-art recommendation techniques.
科研通智能强力驱动
Strongly Powered by AbleSci AI