计算机科学
人工神经网络
人工智能
机器学习
深度学习
特征(语言学)
算法
随机搜索
语言学
哲学
作者
James Bergstra,R. Bardenet,Yoshua Bengio,Balázs Kégl
出处
期刊:Le Centre pour la Communication Scientifique Directe - HAL - Diderot
日期:2016-09-02
被引量:3180
摘要
Several recent advances to the state of the art in image classification benchmarks have come from better configurations of existing techniques rather than novel ap-proaches to feature learning. Traditionally, hyper-parameter optimization has been the job of humans because they can be very efficient in regimes where only a few trials are possible. Presently, computer clusters and GPU processors make it pos-sible to run more trials and we show that algorithmic approaches can find better results. We present hyper-parameter optimization results on tasks of training neu-ral networks and deep belief networks (DBNs). We optimize hyper-parameters using random search and two new greedy sequential methods based on the ex-pected improvement criterion. Random search has been shown to be sufficiently efficient for learning neural networks for several datasets, but we show it is unreli-able for training DBNs. The sequential algorithms are applied to the most difficult DBN learning problems from [1] and find significantly better results than the best previously reported. This work contributes novel techniques for making response surface models P (y|x) in which many elements of hyper-parameter assignment (x) are known to be irrelevant given particular values of other elements. 1
科研通智能强力驱动
Strongly Powered by AbleSci AI