颂歌
解算器
序列(生物学)
黑匣子
连续建模
算法
残余物
常量(计算机编程)
变量(数学)
潜变量
计算机科学
数学
应用数学
微分方程
常微分方程
人工神经网络
人工智能
数学分析
遗传学
生物
程序设计语言
作者
Ricky T. Q. Chen,Yulia Rubanova,Jesse Bettencourt,David Duvenaud
出处
期刊:Cornell University - arXiv
日期:2018-12-03
卷期号:31: 6572-6583
被引量:765
摘要
We introduce a new family of deep neural network models. Instead of specifying a discrete sequence of hidden layers, we parameterize the derivative of the hidden state using a neural network. The output of the network is computed using a black-box differential equation solver. These continuous-depth models have constant memory cost, adapt their evaluation strategy to each input, and can explicitly trade numerical precision for speed. We demonstrate these properties in continuous-depth residual networks and continuous-time latent variable models. We also construct continuous normalizing flows, a generative model that can train by maximum likelihood, without partitioning or ordering the data dimensions. For training, we show how to scalably backpropagate through any ODE solver, without access to its internal operations. This allows end-to-end training of ODEs within larger models.
科研通智能强力驱动
Strongly Powered by AbleSci AI