过度拟合
概化理论
变压器
贝叶斯概率
人工智能
一般化
计算机科学
机器学习
惯性
模式识别(心理学)
工程类
数学
人工神经网络
统计
电压
电气工程
经典力学
物理
数学分析
作者
Yiming Xiao,Haidong Shao,Jie Wang,Shen Yan,Bin Liu
标识
DOI:10.1016/j.ymssp.2023.110936
摘要
Transformer has been widely applied in the research of rotating machinery fault diagnosis due to its ability to explore the internal correlation of vibration signals. However, challenges still exist despite the countless efforts. Generally, Transformer is more prone to overfitting than CNN on small-scale datasets. In practical engineering, collecting sufficient fault samples for training is difficult, resulting in poor generalization of Transformer. In addition, the measured signals are often accompanied with severe noise, further reducing the generalization performance of the model. Meanwhile, the collected signals often follow different distributions due to the changing operating conditions, which places higher demands on the generalizability of Transformer. This paper proposes a Bayesian variational Transformer (Bayesformer) to cope with the abovementioned problems. In Bayesformer, all the attention weights are treated as latent random variables, rather than determined values as the previous studies. This allows to train an ensemble of networks, instead of a single one, enhancing the generalizability of the model. Three experimental studies are conducted to illustrate the developed model and superior diagnostic performance is showed throughout the experiments.
科研通智能强力驱动
Strongly Powered by AbleSci AI