颂歌
人工神经网络
离散化
计算机科学
可分离空间
趋同(经济学)
数学优化
信任域
常微分方程
维数(图论)
功能(生物学)
算法
变量(数学)
人工智能
数学
半径
应用数学
微分方程
数学分析
计算机安全
进化生物学
纯数学
经济
生物
经济增长
作者
Ya-ping Wang,Guangyong Chen,Min Gan
标识
DOI:10.1109/ddcls58216.2023.10167249
摘要
Neural ordinary differential equations (Neural ODE) interprets deep networks as discretization of dynamical systems, and has shown great promise in the physical science, modeling irregular time series, and mean field games. The Neural ODE comsumes a long time training process, which is arguably one of the main stumbling blocks towards their widespread adoption. To improve the convergence speed of training, in this parper, we formulate the training task as a separable nonlinear optimization problem, and propose a separable training algorithm based on a nonmonotone trust-region method. The proposed algorithm uses the variable projection strategy to reduce the dimension of variables by solving a subproblem and then the trust-region method is used to optimize the reduced function. To accelerate the convergence speed, we introduce the nonmonotone strategy to make the update of trust-region radius elastic and employ the adaptive technology that uses the gradient information of the objective function to update the radius. Numerical results confirm the effectiveness of the proposed algorithm.
科研通智能强力驱动
Strongly Powered by AbleSci AI