贝叶斯优化
计算机科学
人工智能
机器学习
超参数
变阶贝叶斯网络
高斯过程
贝叶斯概率
人工神经网络
贝叶斯推理
高斯分布
物理
量子力学
作者
Jost Tobias Springenberg,Aaron Klein,Stefan Falkner,Frank Hutter
出处
期刊:Neural Information Processing Systems
日期:2016-12-05
卷期号:29: 4141-4149
被引量:282
摘要
Bayesian optimization is a prominent method for optimizing expensive-to-evaluate black-box functions that is widely applied to tuning the hyperparameters of machine learning algorithms. Despite its successes, the prototypical Bayesian optimization approach - using Gaussian process models - does not scale well to either many hyperparameters or many function evaluations. Attacking this lack of scalability and flexibility is thus one of the key challenges of the field. We present a general approach for using flexible parametric models (neural networks) for Bayesian optimization, staying as close to a truly Bayesian treatment as possible. We obtain scalability through stochastic gradient Hamiltonian Monte Carlo, whose robustness we improve via a scale adaptation. Experiments including multi-task Bayesian optimization with 21 tasks, parallel optimization of deep neural networks and deep reinforcement learning show the power and flexibility of this approach.
科研通智能强力驱动
Strongly Powered by AbleSci AI