黑森矩阵
遗忘
正规化(语言学)
特征向量
上下界
数学优化
计算机科学
应用数学
基质(化学分析)
人工神经网络
数学
人工智能
数学分析
哲学
语言学
物理
材料科学
量子力学
复合材料
作者
Yajing Kong,Liu Liu,Huanhuan Chen,Janusz Kacprzyk,Dacheng Tao
标识
DOI:10.1109/tnnls.2023.3292359
摘要
Neural networks tend to suffer performance deterioration on previous tasks when they are applied to multiple tasks sequentially without access to previous data. The problem is commonly known as catastrophic forgetting, a significant challenge in continual learning (CL). To overcome the catastrophic forgetting, regularization-based CL methods construct a regularization-based term, which can be considered as the approximation loss function of previous tasks, to penalize the update of parameters. However, the rigorous theoretical analysis of regularization-based methods is limited. Therefore, we theoretically analyze the forgetting and the convergence properties of regularization-based methods. The theoretical results demonstrate that the upper bound of the forgetting has a relationship with the maximum eigenvalue of the Hessian matrix. Hence, to decrease the upper bound of the forgetting, we propose eiGenvalues ExplorAtion Regularization-based (GEAR) method, which explores the geometric properties of the approximation loss of prior tasks regarding the maximum eigenvalue. Extensive experimental results demonstrate that our method mitigates catastrophic forgetting and outperforms existing regularization-based methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI