MNIST数据库
自编码
计算机科学
潜变量
分歧(语言学)
人工智能
后验概率
先验概率
机器学习
潜变量模型
贝叶斯概率
自回归模型
Kullback-Leibler散度
模式识别(心理学)
人工神经网络
数学
计量经济学
哲学
语言学
作者
Jianyong Sun,Fan Song,Qiaohong Li
标识
DOI:10.1016/j.neucom.2023.127103
摘要
Variational recurrent autoencoder model (VRAE) is an appealing technique for capturing the variabilities underlying complex sequential data, which is realized by introducing high-level latent random variables as hidden states. Existing models suffer from the well-known ‘posterior collapse’ problem, meaning that a powerful autoregressive decoder equipped in the model itself could capture all the variabilities and hence leave the latent variables learning nothing from the data. From the perspective of model training, the posterior collapse problem can result in a very low Kullback–Leibler divergence (KL-divergence) value, which means the posteriors of the latent variables tend to be just the priors. In this paper, we address this problem by proposing a Bayesian variational recurrent neural network (BVRNN) model, in which two additional decoders are added into the original VRAE. These extra decoders can force the latent variables to learn meaningful knowledge during the training process. We conduct experiments on MNIST and Fashion-MNIST dataset. The experimental results show that the proposed model outperforms several baseline models. We further adapt the proposed model to a very challenging task in natural language processing, namely Named-Entity Recognition (NER). Experimental results show that our model is competitive to the state-of-the-art models on NER.
科研通智能强力驱动
Strongly Powered by AbleSci AI