计算机科学
火车
光学(聚焦)
GSM演进的增强数据速率
联合学习
分歧(语言学)
班级(哲学)
机器学习
人工智能
数据挖掘
移动设备
人工神经网络
人口
边缘设备
云计算
地理
社会学
人口学
哲学
物理
光学
操作系统
地图学
语言学
作者
Yue Zhao,Meng Li,Liangzhen Lai,Naveen Suda,Damon Civin,Vikas Chandra
出处
期刊:Cornell University - arXiv
日期:2018-01-01
被引量:1511
标识
DOI:10.48550/arxiv.1806.00582
摘要
Federated learning enables resource-constrained edge compute devices, such as mobile phones and IoT devices, to learn a shared model for prediction, while keeping the training data local. This decentralized approach to train models provides privacy, security, regulatory and economic benefits. In this work, we focus on the statistical challenge of federated learning when local data is non-IID. We first show that the accuracy of federated learning reduces significantly, by up to 55% for neural networks trained for highly skewed non-IID data, where each client device trains only on a single class of data. We further show that this accuracy reduction can be explained by the weight divergence, which can be quantified by the earth mover's distance (EMD) between the distribution over classes on each device and the population distribution. As a solution, we propose a strategy to improve training on non-IID data by creating a small subset of data which is globally shared between all the edge devices. Experiments show that accuracy can be increased by 30% for the CIFAR-10 dataset with only 5% globally shared data.
科研通智能强力驱动
Strongly Powered by AbleSci AI