计算机科学
深度学习
人工智能
调试
人工神经网络
深层神经网络
背景(考古学)
机器学习
比例(比率)
培训(气象学)
量子力学
生物
物理
古生物学
气象学
程序设计语言
出处
期刊:Cornell University - arXiv
日期:2012-06-24
被引量:285
摘要
Learning algorithms related to artificial neural networks and in particular for Deep Learning may seem to involve many bells and whistles, called hyper-parameters. This chapter is meant as a practical guide with recommendations for some of the most commonly used hyper-parameters, in particular in the context of learning algorithms based on back-propagated gradient and gradient-based optimization. It also discusses how to deal with the fact that more interesting results can be obtained when allowing one to adjust many hyper-parameters. Overall, it describes elements of the practice used to successfully and efficiently train and debug large-scale and often deep multi-layer neural networks. It closes with open questions about the training difficulties observed with deeper architectures.
科研通智能强力驱动
Strongly Powered by AbleSci AI