计算机科学
可扩展性
卷积神经网络
并行计算
培训(气象学)
趋同(经济学)
任务(项目管理)
平行性(语法)
执行时间
数据并行性
比例(比率)
人工智能
经济
气象学
管理
物理
数据库
量子力学
经济增长
作者
Sunwoo Lee,Qiao Kang,Sandeep Madireddy,Prasanna Balaprakash,Ankit Agrawal,Alok Choudhary,Richard Archibald,Wei‐keng Liao
标识
DOI:10.1109/bigdata47090.2019.9006550
摘要
Training Convolutional Neural Network (CNN) is a computationally intensive task, requiring efficient parallelization to shorten the execution time. Considering the ever-increasing size of available training data, the parallelization of CNN training becomes more important. Data-parallelism, a popular parallelization strategy that distributes the input data among compute processes, requires the mini-batch size to be sufficiently large to achieve a high degree of parallelism. However, training with large batch size is known to produce a low convergence accuracy. In image restoration problems, for example, the batch size is typically tuned to a small value between 16 ~ 64, making it challenging to scale up the training. In this paper, we propose a parallel CNN training strategy that gradually increases the mini-batch size and learning rate at run-time. While improving the scalability, this strategy also maintains the accuracy close to that of the training with a fixed small batch size. We evaluate the performance of the proposed parallel CNN training algorithm with image regression and classification applications using various models and datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI