MNIST数据库
计算机科学
培训(气象学)
生成语法
对抗制
机器学习
过程(计算)
人工智能
点(几何)
功能(生物学)
批处理
选择(遗传算法)
深度学习
数学
物理
几何学
进化生物学
气象学
生物
程序设计语言
操作系统
作者
Bhaskar Ghosh,Indira Kalyan Dutta,Albert Carlson,Michael W. Totaro,Magdy Bayoumi
标识
DOI:10.1109/uemcon51285.2020.9298092
摘要
Increasing the performance of a Generative Adversarial Network (GAN) requires experimentation in choosing the suitable training hyper-parameters of learning rate and batch size. There is no consensus on learning rates or batch sizes in GANs, which makes it a "trial-and-error" process to get acceptable output. Researchers have differing views regarding the effect of batch sizes on run time. This paper investigates the impact of these training parameters of GANs with respect to actual elapsed training time. In our initial experiments, we study the effects of batch sizes, learning rates, loss function, and optimization algorithm on training using the MNIST dataset over 30,000 epochs. The simplicity of the MNIST dataset allows for a starting point in initial studies to understand if the parameter changes have any significant impact on the training times. The goal is to analyze and understand the results of varying loss functions, batch sizes, optimizer algorithms, and learning rates on GANs and address the key issue of batch size and learning rate selection.
科研通智能强力驱动
Strongly Powered by AbleSci AI