维数(图论)
Lasso(编程语言)
梯度下降
计算机科学
线性回归
对数
缩放比例
平滑度
节点(物理)
凸性
算法
数学
数学优化
人工智能
统计
组合数学
人工神经网络
数学分析
结构工程
万维网
金融经济学
工程类
经济
几何学
作者
Ji Yao,Gesualdo Scutari,Ying Sun,Harsha Honnappa
标识
DOI:10.1109/tit.2023.3267742
摘要
We study linear regression from data distributed over a network of agents (with no server node) by means of LASSO estimation, in high-dimension, which allows the ambient dimension to grow faster than the sample size. While there is a vast literature of distributed algorithms applicable to the problem, statistical and computational guarantees of most of them remain unclear in high dimension. This paper provides a first statistical study of the Distributed Gradient Descent (DGD) in the Adapt-Then-Combine (ATC) form. Our theory shows that, under standard notions of restricted strong convexity and smoothness of the loss functions–which hold with high probability for standard data generation models–suitable conditions on the network connectivity and algorithm tuning, DGD-ATC converges globally at a linear rate to an estimate that is within the centralized statistical precision of the model. In the worst-case scenario, the total number of communications to statistical optimality grows logarithmically with the ambient dimension, which improves on the communication complexity of DGD in the Combine-Then-Adapt (CTA) form, scaling linearly with the dimension. This reveals that mixing gradient information among agents, as DGD-ATC does, is critical in high-dimensions to obtain favorable rate scalings.
科研通智能强力驱动
Strongly Powered by AbleSci AI